2026-03-09T13:50:56.116 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-09T13:50:56.121 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T13:50:56.140 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/497 branch: squid description: orch/cephadm/mgr-nfs-upgrade/{0-centos_9.stream 1-bootstrap/17.2.0 1-start 2-nfs 3-upgrade-with-workload 4-final} email: null first_in_suite: false flavor: default job_id: '497' last_in_suite: false machine_type: vps name: kyr-2026-03-09_11:23:05-orch-squid-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_REFRESH_FAILED log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - osd.0 - osd.1 - osd.2 - osd.3 - client.0 - - host.b - osd.4 - osd.5 - osd.6 - osd.7 seed: 3443 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 targets: vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBB9GKtLWXnEpgshtJgIUaygQatKi3zad3UKvJUhP1Gyg1Pof9KAz4aXt9st+4eF95pC/UZG/T1t1CMDlOkF+uko= vm10.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPFIqt0GnGWnMmYLlym2P7ffs3NX9l6aZB6yHYhAW7NMVVT31u42Eqr8XmUBmYR268g/xlNWRGabVdFBVJ4sXFs= tasks: - cephadm: cephadm_branch: v17.2.0 cephadm_git_url: https://github.com/ceph/ceph image: quay.io/ceph/ceph:v17.2.0 roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph fs volume create foofs - cephadm.wait_for_service: service: mds.foofs - cephadm.shell: host.a: - ceph nfs cluster create foo --placement=2 || ceph nfs cluster create cephfs foo --placement=2 - ceph nfs export create cephfs --fsname foofs --clusterid foo --binding /fake || ceph nfs export create cephfs --fsname foofs --cluster-id foo --pseudo-path /fake - while ! ceph orch ls | grep nfs | grep 2/2 ; do sleep 1 ; done - vip.exec: host.a: - mkdir /mnt/foo - while ! mount -t nfs $(hostname):/fake /mnt/foo -o sync ; do sleep 5 ; done - echo test > /mnt/foo/testfile - sync - parallel: - upgrade-tasks - workload-tasks - vip.exec: host.a: - umount /mnt/foo - cephadm.shell: host.a: - ceph nfs cluster ls | grep foo - ceph nfs export ls foo --detailed - rados -p .nfs --all ls - - cephadm.shell: host.a: - 'set -ex [[ `ceph config get mgr mgr/cephadm/migration_current` -gt 2 ]] ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-09_11:23:05 tube: vps upgrade-tasks: sequential: - cephadm.shell: env: - sha1 host.a: - ceph config set mon mon_warn_on_insecure_global_id_reclaim false --force - ceph config set mon mon_warn_on_insecure_global_id_reclaim_allowed false --force - ceph config set global log_to_journald false --force - ceph mgr module enable nfs --force - ceph orch upgrade start --image quay.ceph.io/ceph-ci/ceph:$sha1 - cephadm.shell: env: - sha1 host.a: - while ceph orch upgrade status | jq '.in_progress' | grep true && ! ceph orch upgrade status | jq '.message' | grep Error ; do ceph orch ps ; ceph versions ; ceph orch upgrade status ; ceph health detail ; sleep 30 ; done - ceph orch ps - ceph versions - echo "wait for servicemap items w/ changing names to refresh" - sleep 60 - ceph orch ps - ceph orch upgrade status - ceph health detail - ceph versions - ceph versions | jq -e '.overall | length == 1' - ceph versions | jq -e '.overall | keys' | grep $sha1 - cephadm.wait_for_service: service: nfs.foo user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 workload-tasks: sequential: - exec: host.a: - cd /mnt/foo && dbench 5 -t 600 || true - umount /mnt/foo - while ! mount -t nfs $(hostname):/fake /mnt/foo ; do sleep 5 ; done - cd /mnt/foo && dbench 5 -t 5 2026-03-09T13:50:56.140 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa; will attempt to use it 2026-03-09T13:50:56.140 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa/tasks 2026-03-09T13:50:56.140 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-09T13:50:56.141 INFO:teuthology.task.internal:Checking packages... 2026-03-09T13:50:56.141 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-09T13:50:56.141 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-09T13:50:56.141 INFO:teuthology.packaging:ref: None 2026-03-09T13:50:56.141 INFO:teuthology.packaging:tag: None 2026-03-09T13:50:56.141 INFO:teuthology.packaging:branch: squid 2026-03-09T13:50:56.141 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T13:50:56.141 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-09T13:50:56.920 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-09T13:50:56.921 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-09T13:50:56.922 INFO:teuthology.task.internal:no buildpackages task found 2026-03-09T13:50:56.922 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-09T13:50:56.922 INFO:teuthology.task.internal:Saving configuration 2026-03-09T13:50:56.928 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-09T13:50:56.930 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-09T13:50:56.936 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm08.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/497', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 13:49:24.354200', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:08', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBB9GKtLWXnEpgshtJgIUaygQatKi3zad3UKvJUhP1Gyg1Pof9KAz4aXt9st+4eF95pC/UZG/T1t1CMDlOkF+uko='} 2026-03-09T13:50:56.942 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm10.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/497', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 13:49:24.354831', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:0a', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPFIqt0GnGWnMmYLlym2P7ffs3NX9l6aZB6yHYhAW7NMVVT31u42Eqr8XmUBmYR268g/xlNWRGabVdFBVJ4sXFs='} 2026-03-09T13:50:56.942 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-09T13:50:56.943 INFO:teuthology.task.internal:roles: ubuntu@vm08.local - ['host.a', 'osd.0', 'osd.1', 'osd.2', 'osd.3', 'client.0'] 2026-03-09T13:50:56.943 INFO:teuthology.task.internal:roles: ubuntu@vm10.local - ['host.b', 'osd.4', 'osd.5', 'osd.6', 'osd.7'] 2026-03-09T13:50:56.943 INFO:teuthology.run_tasks:Running task console_log... 2026-03-09T13:50:56.950 DEBUG:teuthology.task.console_log:vm08 does not support IPMI; excluding 2026-03-09T13:50:56.956 DEBUG:teuthology.task.console_log:vm10 does not support IPMI; excluding 2026-03-09T13:50:56.957 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f56a5a76170>, signals=[15]) 2026-03-09T13:50:56.957 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-09T13:50:56.958 INFO:teuthology.task.internal:Opening connections... 2026-03-09T13:50:56.958 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-03-09T13:50:56.958 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T13:50:57.015 DEBUG:teuthology.task.internal:connecting to ubuntu@vm10.local 2026-03-09T13:50:57.016 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm10.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T13:50:57.074 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-09T13:50:57.075 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-03-09T13:50:57.120 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-03-09T13:50:57.120 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-03-09T13:50:57.174 INFO:teuthology.orchestra.run.vm08.stdout:NAME="CentOS Stream" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="9" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:ID="centos" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE="rhel fedora" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="9" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:PLATFORM_ID="platform:el9" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:ANSI_COLOR="0;31" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:LOGO="fedora-logo-icon" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://centos.org/" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T13:50:57.175 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T13:50:57.175 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-03-09T13:50:57.181 DEBUG:teuthology.orchestra.run.vm10:> uname -m 2026-03-09T13:50:57.196 INFO:teuthology.orchestra.run.vm10.stdout:x86_64 2026-03-09T13:50:57.196 DEBUG:teuthology.orchestra.run.vm10:> cat /etc/os-release 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:NAME="CentOS Stream" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:VERSION="9" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:ID="centos" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:ID_LIKE="rhel fedora" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:VERSION_ID="9" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:PLATFORM_ID="platform:el9" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:ANSI_COLOR="0;31" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:LOGO="fedora-logo-icon" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:HOME_URL="https://centos.org/" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T13:50:57.252 INFO:teuthology.orchestra.run.vm10.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T13:50:57.252 INFO:teuthology.lock.ops:Updating vm10.local on lock server 2026-03-09T13:50:57.258 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-09T13:50:57.260 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-09T13:50:57.260 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-09T13:50:57.260 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-03-09T13:50:57.262 DEBUG:teuthology.orchestra.run.vm10:> test '!' -e /home/ubuntu/cephtest 2026-03-09T13:50:57.306 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-09T13:50:57.308 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-09T13:50:57.308 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-03-09T13:50:57.319 DEBUG:teuthology.orchestra.run.vm10:> test -z $(ls -A /var/lib/ceph) 2026-03-09T13:50:57.333 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T13:50:57.366 INFO:teuthology.orchestra.run.vm10.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T13:50:57.366 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-09T13:50:57.375 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-03-09T13:50:57.391 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T13:50:57.587 DEBUG:teuthology.orchestra.run.vm10:> test -e /ceph-qa-ready 2026-03-09T13:50:57.602 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T13:50:57.796 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-09T13:50:57.797 INFO:teuthology.task.internal:Creating test directory... 2026-03-09T13:50:57.798 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T13:50:57.800 DEBUG:teuthology.orchestra.run.vm10:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T13:50:57.818 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-09T13:50:57.819 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-09T13:50:57.821 INFO:teuthology.task.internal:Creating archive directory... 2026-03-09T13:50:57.821 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T13:50:57.859 DEBUG:teuthology.orchestra.run.vm10:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T13:50:57.879 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-09T13:50:57.880 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-09T13:50:57.880 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T13:50:57.934 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T13:50:57.934 DEBUG:teuthology.orchestra.run.vm10:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T13:50:57.951 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T13:50:57.951 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T13:50:57.976 DEBUG:teuthology.orchestra.run.vm10:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T13:50:57.999 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T13:50:58.008 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T13:50:58.019 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T13:50:58.029 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T13:50:58.031 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-09T13:50:58.032 INFO:teuthology.task.internal:Configuring sudo... 2026-03-09T13:50:58.032 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T13:50:58.052 DEBUG:teuthology.orchestra.run.vm10:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T13:50:58.098 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-09T13:50:58.101 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-09T13:50:58.101 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T13:50:58.117 DEBUG:teuthology.orchestra.run.vm10:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T13:50:58.155 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T13:50:58.194 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T13:50:58.250 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:50:58.250 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T13:50:58.307 DEBUG:teuthology.orchestra.run.vm10:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T13:50:58.333 DEBUG:teuthology.orchestra.run.vm10:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T13:50:58.390 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T13:50:58.390 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T13:50:58.452 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-03-09T13:50:58.454 DEBUG:teuthology.orchestra.run.vm10:> sudo service rsyslog restart 2026-03-09T13:50:58.478 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T13:50:58.523 INFO:teuthology.orchestra.run.vm10.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T13:50:58.972 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-09T13:50:58.973 INFO:teuthology.task.internal:Starting timer... 2026-03-09T13:50:58.973 INFO:teuthology.run_tasks:Running task pcp... 2026-03-09T13:50:58.977 INFO:teuthology.run_tasks:Running task selinux... 2026-03-09T13:50:58.979 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-09T13:50:58.979 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-03-09T13:50:58.979 INFO:teuthology.task.selinux:Excluding vm10: VMs are not yet supported 2026-03-09T13:50:58.979 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-09T13:50:58.979 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-09T13:50:58.979 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-09T13:50:58.979 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-09T13:50:58.980 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-09T13:50:58.981 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-09T13:50:58.982 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-09T13:50:59.655 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-09T13:50:59.661 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-09T13:50:59.661 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryv10n3dn_ --limit vm08.local,vm10.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-09T13:53:01.752 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm08.local'), Remote(name='ubuntu@vm10.local')] 2026-03-09T13:53:01.753 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-03-09T13:53:01.753 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T13:53:01.820 DEBUG:teuthology.orchestra.run.vm08:> true 2026-03-09T13:53:01.901 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-03-09T13:53:01.901 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm10.local' 2026-03-09T13:53:01.901 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm10.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T13:53:01.969 DEBUG:teuthology.orchestra.run.vm10:> true 2026-03-09T13:53:02.043 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm10.local' 2026-03-09T13:53:02.043 INFO:teuthology.run_tasks:Running task clock... 2026-03-09T13:53:02.047 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-09T13:53:02.047 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T13:53:02.047 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T13:53:02.049 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T13:53:02.049 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T13:53:02.088 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T13:53:02.103 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T13:53:02.121 INFO:teuthology.orchestra.run.vm10.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T13:53:02.131 INFO:teuthology.orchestra.run.vm08.stderr:sudo: ntpd: command not found 2026-03-09T13:53:02.140 INFO:teuthology.orchestra.run.vm10.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T13:53:02.144 INFO:teuthology.orchestra.run.vm08.stdout:506 Cannot talk to daemon 2026-03-09T13:53:02.162 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T13:53:02.175 INFO:teuthology.orchestra.run.vm10.stderr:sudo: ntpd: command not found 2026-03-09T13:53:02.178 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T13:53:02.191 INFO:teuthology.orchestra.run.vm10.stdout:506 Cannot talk to daemon 2026-03-09T13:53:02.215 INFO:teuthology.orchestra.run.vm10.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T13:53:02.230 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-03-09T13:53:02.233 INFO:teuthology.orchestra.run.vm10.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T13:53:02.233 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T13:53:02.233 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-03-09T13:53:02.292 INFO:teuthology.orchestra.run.vm10.stderr:bash: line 1: ntpq: command not found 2026-03-09T13:53:02.294 INFO:teuthology.orchestra.run.vm10.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T13:53:02.294 INFO:teuthology.orchestra.run.vm10.stdout:=============================================================================== 2026-03-09T13:53:02.294 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-09T13:53:02.350 INFO:tasks.cephadm:Config: {'cephadm_branch': 'v17.2.0', 'cephadm_git_url': 'https://github.com/ceph/ceph', 'image': 'quay.io/ceph/ceph:v17.2.0', 'roleless': True, 'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_REFRESH_FAILED'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-09T13:53:02.350 INFO:tasks.cephadm:Cluster image is quay.io/ceph/ceph:v17.2.0 2026-03-09T13:53:02.350 INFO:tasks.cephadm:Cluster fsid is 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:02.350 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-09T13:53:02.350 INFO:tasks.cephadm:No mon roles; fabricating mons 2026-03-09T13:53:02.350 INFO:tasks.cephadm:Monitor IPs: {'mon.vm08': '192.168.123.108', 'mon.vm10': '192.168.123.110'} 2026-03-09T13:53:02.350 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-09T13:53:02.350 DEBUG:teuthology.orchestra.run.vm08:> sudo hostname $(hostname -s) 2026-03-09T13:53:02.384 DEBUG:teuthology.orchestra.run.vm10:> sudo hostname $(hostname -s) 2026-03-09T13:53:02.425 INFO:tasks.cephadm:Downloading cephadm (repo https://github.com/ceph/ceph ref v17.2.0)... 2026-03-09T13:53:02.425 DEBUG:teuthology.orchestra.run.vm08:> curl --silent https://raw.githubusercontent.com/ceph/ceph/v17.2.0/src/cephadm/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T13:53:02.727 INFO:teuthology.orchestra.run.vm08.stdout:-rw-r--r--. 1 ubuntu ubuntu 320521 Mar 9 13:53 /home/ubuntu/cephtest/cephadm 2026-03-09T13:53:02.728 DEBUG:teuthology.orchestra.run.vm10:> curl --silent https://raw.githubusercontent.com/ceph/ceph/v17.2.0/src/cephadm/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T13:53:02.816 INFO:teuthology.orchestra.run.vm10.stdout:-rw-r--r--. 1 ubuntu ubuntu 320521 Mar 9 13:53 /home/ubuntu/cephtest/cephadm 2026-03-09T13:53:02.816 DEBUG:teuthology.orchestra.run.vm08:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T13:53:02.837 DEBUG:teuthology.orchestra.run.vm10:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T13:53:02.861 INFO:tasks.cephadm:Pulling image quay.io/ceph/ceph:v17.2.0 on all hosts... 2026-03-09T13:53:02.862 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 pull 2026-03-09T13:53:02.878 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 pull 2026-03-09T13:53:03.081 INFO:teuthology.orchestra.run.vm08.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-09T13:53:03.136 INFO:teuthology.orchestra.run.vm10.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-09T13:53:28.244 INFO:teuthology.orchestra.run.vm08.stdout:{ 2026-03-09T13:53:28.244 INFO:teuthology.orchestra.run.vm08.stdout: "ceph_version": "ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable)", 2026-03-09T13:53:28.244 INFO:teuthology.orchestra.run.vm08.stdout: "image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", 2026-03-09T13:53:28.244 INFO:teuthology.orchestra.run.vm08.stdout: "repo_digests": [ 2026-03-09T13:53:28.244 INFO:teuthology.orchestra.run.vm08.stdout: "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", 2026-03-09T13:53:28.244 INFO:teuthology.orchestra.run.vm08.stdout: "quay.io/ceph/ceph@sha256:cb4d698cb769b6aba05bf6ef04f41a7fe694160140347576e13bd9348514b667" 2026-03-09T13:53:28.244 INFO:teuthology.orchestra.run.vm08.stdout: ] 2026-03-09T13:53:28.244 INFO:teuthology.orchestra.run.vm08.stdout:} 2026-03-09T13:53:28.596 INFO:teuthology.orchestra.run.vm10.stdout:{ 2026-03-09T13:53:28.596 INFO:teuthology.orchestra.run.vm10.stdout: "ceph_version": "ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable)", 2026-03-09T13:53:28.596 INFO:teuthology.orchestra.run.vm10.stdout: "image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", 2026-03-09T13:53:28.596 INFO:teuthology.orchestra.run.vm10.stdout: "repo_digests": [ 2026-03-09T13:53:28.596 INFO:teuthology.orchestra.run.vm10.stdout: "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", 2026-03-09T13:53:28.596 INFO:teuthology.orchestra.run.vm10.stdout: "quay.io/ceph/ceph@sha256:cb4d698cb769b6aba05bf6ef04f41a7fe694160140347576e13bd9348514b667" 2026-03-09T13:53:28.596 INFO:teuthology.orchestra.run.vm10.stdout: ] 2026-03-09T13:53:28.596 INFO:teuthology.orchestra.run.vm10.stdout:} 2026-03-09T13:53:28.618 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph 2026-03-09T13:53:28.653 DEBUG:teuthology.orchestra.run.vm10:> sudo mkdir -p /etc/ceph 2026-03-09T13:53:28.692 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /etc/ceph 2026-03-09T13:53:28.726 DEBUG:teuthology.orchestra.run.vm10:> sudo chmod 777 /etc/ceph 2026-03-09T13:53:28.764 INFO:tasks.cephadm:Writing seed config... 2026-03-09T13:53:28.764 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-09T13:53:28.764 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-09T13:53:28.764 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-09T13:53:28.764 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-09T13:53:28.764 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-09T13:53:28.764 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-09T13:53:28.764 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-09T13:53:28.764 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-09T13:53:28.765 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-09T13:53:28.765 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:53:28.765 DEBUG:teuthology.orchestra.run.vm08:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-09T13:53:28.795 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-09T13:53:28.795 DEBUG:teuthology.orchestra.run.vm08:mon.vm08> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mon.vm08.service 2026-03-09T13:53:28.840 INFO:tasks.cephadm:Bootstrapping... 2026-03-09T13:53:28.840 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 -v bootstrap --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-ip 192.168.123.108 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-09T13:53:29.003 INFO:teuthology.orchestra.run.vm08.stderr:-------------------------------------------------------------------------------- 2026-03-09T13:53:29.004 INFO:teuthology.orchestra.run.vm08.stderr:cephadm ['--image', 'quay.io/ceph/ceph:v17.2.0', '-v', 'bootstrap', '--fsid', '4a3e67a6-1bbf-11f1-aca1-21c5f680cc64', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-ip', '192.168.123.108', '--skip-admin-label'] 2026-03-09T13:53:29.025 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: 5.8.0 2026-03-09T13:53:29.031 INFO:teuthology.orchestra.run.vm08.stderr:Verifying podman|docker is present... 2026-03-09T13:53:29.051 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: 5.8.0 2026-03-09T13:53:29.055 INFO:teuthology.orchestra.run.vm08.stderr:Verifying lvm2 is present... 2026-03-09T13:53:29.055 INFO:teuthology.orchestra.run.vm08.stderr:Verifying time synchronization is in place... 2026-03-09T13:53:29.061 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T13:53:29.068 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: inactive 2026-03-09T13:53:29.076 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: enabled 2026-03-09T13:53:29.083 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: active 2026-03-09T13:53:29.084 INFO:teuthology.orchestra.run.vm08.stderr:Unit chronyd.service is enabled and running 2026-03-09T13:53:29.084 INFO:teuthology.orchestra.run.vm08.stderr:Repeating the final host check... 2026-03-09T13:53:29.102 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: 5.8.0 2026-03-09T13:53:29.106 INFO:teuthology.orchestra.run.vm08.stderr:podman (/bin/podman) version 5.8.0 is present 2026-03-09T13:53:29.106 INFO:teuthology.orchestra.run.vm08.stderr:systemctl is present 2026-03-09T13:53:29.106 INFO:teuthology.orchestra.run.vm08.stderr:lvcreate is present 2026-03-09T13:53:29.112 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T13:53:29.120 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: inactive 2026-03-09T13:53:29.134 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: enabled 2026-03-09T13:53:29.142 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: active 2026-03-09T13:53:29.143 INFO:teuthology.orchestra.run.vm08.stderr:Unit chronyd.service is enabled and running 2026-03-09T13:53:29.143 INFO:teuthology.orchestra.run.vm08.stderr:Host looks OK 2026-03-09T13:53:29.143 INFO:teuthology.orchestra.run.vm08.stderr:Cluster fsid: 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:29.143 INFO:teuthology.orchestra.run.vm08.stderr:Acquiring lock 140327122256032 on /run/cephadm/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64.lock 2026-03-09T13:53:29.144 INFO:teuthology.orchestra.run.vm08.stderr:Lock 140327122256032 acquired on /run/cephadm/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64.lock 2026-03-09T13:53:29.144 INFO:teuthology.orchestra.run.vm08.stderr:Verifying IP 192.168.123.108 port 3300 ... 2026-03-09T13:53:29.144 INFO:teuthology.orchestra.run.vm08.stderr:Verifying IP 192.168.123.108 port 6789 ... 2026-03-09T13:53:29.144 INFO:teuthology.orchestra.run.vm08.stderr:Base mon IP is 192.168.123.108, final addrv is [v2:192.168.123.108:3300,v1:192.168.123.108:6789] 2026-03-09T13:53:29.149 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.108 metric 100 2026-03-09T13:53:29.149 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.108 metric 100 2026-03-09T13:53:29.153 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: ::1 dev lo proto kernel metric 256 pref medium 2026-03-09T13:53:29.153 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-09T13:53:29.156 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-09T13:53:29.157 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: inet6 ::1/128 scope host 2026-03-09T13:53:29.157 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: valid_lft forever preferred_lft forever 2026-03-09T13:53:29.157 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: 2: eth0: mtu 1500 state UP qlen 1000 2026-03-09T13:53:29.157 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: inet6 fe80::5055:ff:fe00:8/64 scope link noprefixroute 2026-03-09T13:53:29.157 INFO:teuthology.orchestra.run.vm08.stderr:/sbin/ip: valid_lft forever preferred_lft forever 2026-03-09T13:53:29.157 INFO:teuthology.orchestra.run.vm08.stderr:Mon IP `192.168.123.108` is in CIDR network `192.168.123.0/24` 2026-03-09T13:53:29.158 INFO:teuthology.orchestra.run.vm08.stderr:- internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-09T13:53:29.158 INFO:teuthology.orchestra.run.vm08.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-09T13:53:29.181 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Trying to pull quay.io/ceph/ceph:v17.2.0... 2026-03-09T13:53:30.458 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Getting image source signatures 2026-03-09T13:53:30.458 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Copying blob sha256:89b4a75bc2d8500f15463747507c9623df43886c134463e7f0527e70900e7a7b 2026-03-09T13:53:30.458 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Copying blob sha256:33ca8fff7868c4dc0c11e09bca97c720eb9cfbab7221216754367dd8de70388a 2026-03-09T13:53:30.459 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Copying blob sha256:c32ab78b488d0c72f64eded765c0cf6b5bf2c75dab66cb62a9d367fa6ec42513 2026-03-09T13:53:30.459 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Copying blob sha256:a70843738bb77e1ab9c1f85969ebdfa55f178e746be081d1cb4f94011f69eb7c 2026-03-09T13:53:30.459 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Copying blob sha256:599d07cb321ff0a3c82224e1138fc685793fa69b93ed5780415751a5f7e4b8c2 2026-03-09T13:53:30.459 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Copying config sha256:e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9 2026-03-09T13:53:30.461 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: Writing manifest to image destination 2026-03-09T13:53:30.465 INFO:teuthology.orchestra.run.vm08.stderr:/bin/podman: e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9 2026-03-09T13:53:30.609 INFO:teuthology.orchestra.run.vm08.stderr:ceph: ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable) 2026-03-09T13:53:30.672 INFO:teuthology.orchestra.run.vm08.stderr:Ceph version: ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable) 2026-03-09T13:53:30.672 INFO:teuthology.orchestra.run.vm08.stderr:Extracting ceph user uid/gid from container image... 2026-03-09T13:53:30.760 INFO:teuthology.orchestra.run.vm08.stderr:stat: 167 167 2026-03-09T13:53:30.793 INFO:teuthology.orchestra.run.vm08.stderr:Creating initial keys... 2026-03-09T13:53:30.885 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph-authtool: AQDa0K5pD2SoNBAA9Svdde0SGnUlqw9NGzK9uQ== 2026-03-09T13:53:31.041 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph-authtool: AQDb0K5pMTVkAhAAn2yiilqpCmU07oeJHIfU4A== 2026-03-09T13:53:31.177 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph-authtool: AQDb0K5pDtmIChAAmubQvvYM6al5hTme9WJ+fg== 2026-03-09T13:53:31.204 INFO:teuthology.orchestra.run.vm08.stderr:Creating initial monmap... 2026-03-09T13:53:31.291 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T13:53:31.291 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/monmaptool: setting min_mon_release = octopus 2026-03-09T13:53:31.291 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: set fsid to 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:31.292 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T13:53:31.311 INFO:teuthology.orchestra.run.vm08.stderr:monmaptool for vm08 [v2:192.168.123.108:3300,v1:192.168.123.108:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T13:53:31.311 INFO:teuthology.orchestra.run.vm08.stderr:setting min_mon_release = octopus 2026-03-09T13:53:31.311 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/monmaptool: set fsid to 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:31.311 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T13:53:31.311 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:53:31.311 INFO:teuthology.orchestra.run.vm08.stderr:Creating mon... 2026-03-09T13:53:31.501 INFO:teuthology.orchestra.run.vm08.stderr:create mon.vm08 on 2026-03-09T13:53:31.673 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-09T13:53:31.839 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64.target → /etc/systemd/system/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64.target. 2026-03-09T13:53:31.840 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Created symlink /etc/systemd/system/ceph.target.wants/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64.target → /etc/systemd/system/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64.target. 2026-03-09T13:53:32.212 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Failed to reset failed state of unit ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mon.vm08.service: Unit ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mon.vm08.service not loaded. 2026-03-09T13:53:32.222 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Created symlink /etc/systemd/system/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64.target.wants/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mon.vm08.service → /etc/systemd/system/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@.service. 2026-03-09T13:53:32.592 INFO:teuthology.orchestra.run.vm08.stderr:firewalld does not appear to be present 2026-03-09T13:53:32.592 INFO:teuthology.orchestra.run.vm08.stderr:Not possible to enable service . firewalld.service is not available 2026-03-09T13:53:32.592 INFO:teuthology.orchestra.run.vm08.stderr:Waiting for mon to start... 2026-03-09T13:53:32.592 INFO:teuthology.orchestra.run.vm08.stderr:Waiting for mon... 2026-03-09T13:53:32.672 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:32 vm08 ceph-mon[47466]: mkfs 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: cluster: 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: id: 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: health: HEALTH_OK 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: services: 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mon: 1 daemons, quorum vm08 (age 0.157312s) 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mgr: no daemons active 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd: 0 osds: 0 up, 0 in 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: data: 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: pools: 0 pools, 0 pgs 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: objects: 0 objects, 0 B 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: usage: 0 B used, 0 B / 0 B avail 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: pgs: 2026-03-09T13:53:32.818 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:32.852 INFO:teuthology.orchestra.run.vm08.stderr:mon is available 2026-03-09T13:53:32.852 INFO:teuthology.orchestra.run.vm08.stderr:Assimilating anything we can from ceph.conf... 2026-03-09T13:53:32.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:32 vm08 ceph-mon[47466]: mon.vm08 is new leader, mons vm08 in quorum (ranks 0) 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: [global] 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: fsid = 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mon_host = [v2:192.168.123.108:3300,v1:192.168.123.108:6789] 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mon_osd_allow_pg_remap = true 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mon_osd_allow_primary_affinity = true 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mon_warn_on_no_sortbitwise = false 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd_crush_chooseleaf_type = 0 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: [mgr] 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mgr/telemetry/nag = false 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: [osd] 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd_map_max_advance = 10 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd_mclock_iops_capacity_threshold_hdd = 49000 2026-03-09T13:53:33.069 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd_sloppy_crc = true 2026-03-09T13:53:33.739 INFO:teuthology.orchestra.run.vm08.stderr:Generating new minimal ceph.conf... 2026-03-09T13:53:33.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:33 vm08 ceph-mon[47466]: mon.vm08 is new leader, mons vm08 in quorum (ranks 0) 2026-03-09T13:53:33.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:33 vm08 ceph-mon[47466]: monmap e1: 1 mons at {vm08=[v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0]} 2026-03-09T13:53:33.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:33 vm08 ceph-mon[47466]: fsmap 2026-03-09T13:53:33.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:33 vm08 ceph-mon[47466]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T13:53:33.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:33 vm08 ceph-mon[47466]: mgrmap e1: no daemons active 2026-03-09T13:53:33.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:33 vm08 ceph-mon[47466]: from='client.? 192.168.123.108:0/3152625382' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-09T13:53:33.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:33 vm08 ceph-mon[47466]: from='client.? 192.168.123.108:0/63477154' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T13:53:33.928 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:33 vm08 ceph-mon[47466]: from='client.? 192.168.123.108:0/63477154' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-09T13:53:33.993 INFO:teuthology.orchestra.run.vm08.stderr:Restarting the monitor... 2026-03-09T13:53:34.179 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 systemd[1]: Stopping Ceph mon.vm08 for 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64... 2026-03-09T13:53:34.179 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 bash[47654]: Error: no container with name or ID "ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon.vm08" found: no such container 2026-03-09T13:53:34.179 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08[47462]: 2026-03-09T13:53:34.119+0000 7f559538b700 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.vm08 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T13:53:34.179 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08[47462]: 2026-03-09T13:53:34.119+0000 7f559538b700 -1 mon.vm08@0(leader) e1 *** Got Signal Terminated *** 2026-03-09T13:53:34.179 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 podman[47661]: 2026-03-09 13:53:34.139209936 +0000 UTC m=+0.090816163 container died d23444e00d2b990cb032a6dddb6f013280e8d4f368271554f9818fc829f9122b (image=quay.io/ceph/ceph:v17.2.0, name=ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08, com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, com.redhat.component=centos-stream-container, com.redhat.license_terms=https://centos.org/legal/licensing-policy/, io.buildah.version=1.19.8, io.k8s.display-name=CentOS Stream 8, vcs-type=git, GIT_BRANCH=HEAD, ceph=True, release=754, GIT_REPO=https://github.com/ceph/ceph-container.git, distribution-scope=public, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb, vendor=Red Hat, Inc., version=8, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-754, io.openshift.tags=base centos centos-stream, maintainer=Guillaume Abrioux , GIT_CLEAN=True, GIT_COMMIT=b613db0f44179c0940781c1c7fe04e1acb7093ac, RELEASE=HEAD, architecture=x86_64, build-date=2022-05-03T08:36:31.336870, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, io.openshift.expose-services=, CEPH_POINT_RELEASE=-17.2.0, name=centos-stream) 2026-03-09T13:53:34.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 podman[47661]: 2026-03-09 13:53:34.233214377 +0000 UTC m=+0.184820594 container remove d23444e00d2b990cb032a6dddb6f013280e8d4f368271554f9818fc829f9122b (image=quay.io/ceph/ceph:v17.2.0, name=ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08, CEPH_POINT_RELEASE=-17.2.0, GIT_BRANCH=HEAD, io.openshift.tags=base centos centos-stream, maintainer=Guillaume Abrioux , architecture=x86_64, release=754, RELEASE=HEAD, name=centos-stream, com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, com.redhat.component=centos-stream-container, GIT_REPO=https://github.com/ceph/ceph-container.git, io.buildah.version=1.19.8, io.k8s.display-name=CentOS Stream 8, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image, GIT_COMMIT=b613db0f44179c0940781c1c7fe04e1acb7093ac, ceph=True, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.openshift.expose-services=, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-754, com.redhat.license_terms=https://centos.org/legal/licensing-policy/, vcs-type=git, vendor=Red Hat, Inc., version=8, distribution-scope=public, GIT_CLEAN=True, build-date=2022-05-03T08:36:31.336870) 2026-03-09T13:53:34.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 bash[47661]: ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08 2026-03-09T13:53:34.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 bash[47680]: Error: no container with name or ID "ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon.vm08" found: no such container 2026-03-09T13:53:34.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 systemd[1]: ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mon.vm08.service: Deactivated successfully. 2026-03-09T13:53:34.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 systemd[1]: Stopped Ceph mon.vm08 for 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64. 2026-03-09T13:53:34.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 systemd[1]: Starting Ceph mon.vm08 for 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64... 2026-03-09T13:53:34.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 podman[47719]: 2026-03-09 13:53:34.390509741 +0000 UTC m=+0.025481055 container create b2deb92311981ef3a89c5fa85e3d33c491652f50a682118a54bce428cacbf283 (image=quay.io/ceph/ceph:v17.2.0, name=ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., GIT_CLEAN=True, GIT_COMMIT=b613db0f44179c0940781c1c7fe04e1acb7093ac, maintainer=Guillaume Abrioux , com.redhat.license_terms=https://centos.org/legal/licensing-policy/, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, distribution-scope=public, com.redhat.component=centos-stream-container, CEPH_POINT_RELEASE=-17.2.0, io.buildah.version=1.19.8, name=centos-stream, build-date=2022-05-03T08:36:31.336870, GIT_REPO=https://github.com/ceph/ceph-container.git, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image, version=8, io.k8s.display-name=CentOS Stream 8, GIT_BRANCH=HEAD, release=754, ceph=True, io.openshift.expose-services=, RELEASE=HEAD, vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb, vcs-type=git, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-754, vendor=Red Hat, Inc., architecture=x86_64, io.openshift.tags=base centos centos-stream) 2026-03-09T13:53:34.432 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 podman[47719]: 2026-03-09 13:53:34.431640806 +0000 UTC m=+0.066612130 container init b2deb92311981ef3a89c5fa85e3d33c491652f50a682118a54bce428cacbf283 (image=quay.io/ceph/ceph:v17.2.0, name=ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08, version=8, io.buildah.version=1.19.8, com.redhat.license_terms=https://centos.org/legal/licensing-policy/, io.openshift.expose-services=, build-date=2022-05-03T08:36:31.336870, maintainer=Guillaume Abrioux , com.redhat.component=centos-stream-container, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-754, CEPH_POINT_RELEASE=-17.2.0, architecture=x86_64, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, RELEASE=HEAD, GIT_REPO=https://github.com/ceph/ceph-container.git, vendor=Red Hat, Inc., name=centos-stream, io.openshift.tags=base centos centos-stream, GIT_BRANCH=HEAD, ceph=True, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb, com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, io.k8s.display-name=CentOS Stream 8, distribution-scope=public, vcs-type=git, release=754, GIT_CLEAN=True, GIT_COMMIT=b613db0f44179c0940781c1c7fe04e1acb7093ac, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image) 2026-03-09T13:53:34.452 INFO:teuthology.orchestra.run.vm08.stderr:Setting mon public_network to 192.168.123.0/24 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 podman[47719]: 2026-03-09 13:53:34.43766463 +0000 UTC m=+0.072635944 container start b2deb92311981ef3a89c5fa85e3d33c491652f50a682118a54bce428cacbf283 (image=quay.io/ceph/ceph:v17.2.0, name=ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08, io.k8s.display-name=CentOS Stream 8, distribution-scope=public, ceph=True, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., RELEASE=HEAD, build-date=2022-05-03T08:36:31.336870, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image, version=8, vendor=Red Hat, Inc., GIT_BRANCH=HEAD, release=754, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, io.openshift.expose-services=, GIT_CLEAN=True, maintainer=Guillaume Abrioux , com.redhat.component=centos-stream-container, CEPH_POINT_RELEASE=-17.2.0, vcs-type=git, GIT_COMMIT=b613db0f44179c0940781c1c7fe04e1acb7093ac, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-754, io.buildah.version=1.19.8, architecture=x86_64, name=centos-stream, io.openshift.tags=base centos centos-stream, com.redhat.license_terms=https://centos.org/legal/licensing-policy/, vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb) 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 bash[47719]: b2deb92311981ef3a89c5fa85e3d33c491652f50a682118a54bce428cacbf283 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 podman[47719]: 2026-03-09 13:53:34.378757624 +0000 UTC m=+0.013728938 image pull e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9 quay.io/ceph/ceph:v17.2.0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 systemd[1]: Started Ceph mon.vm08 for 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64. 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: set uid:gid to 167:167 (ceph:ceph) 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable), process ceph-mon, pid 2 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: pidfile_write: ignore empty --pid-file 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: load: jerasure load: lrc 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: RocksDB version: 6.15.5 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Git sha rocksdb_build_git_sha:@0@ 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Compile date Apr 18 2022 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: DB SUMMARY 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: DB Session ID: YB7R1PJK8OJ2G281WZFI 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: CURRENT file: CURRENT 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: IDENTITY file: IDENTITY 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: MANIFEST file: MANIFEST-000009 size: 131 Bytes 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: SST files in /var/lib/ceph/mon/ceph-vm08/store.db dir, Total Num: 1, files: 000008.sst 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm08/store.db: 000010.log size: 80577 ; 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.error_if_exists: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.create_if_missing: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.paranoid_checks: 1 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.env: 0x55a79e4c6860 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.fs: Posix File System 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.info_log: 0x55a7a08f9ee0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_file_opening_threads: 16 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.statistics: (nil) 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.use_fsync: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_log_file_size: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.keep_log_file_num: 1000 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.recycle_log_file_num: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.allow_fallocate: 1 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.allow_mmap_reads: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.allow_mmap_writes: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.use_direct_reads: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.create_missing_column_families: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.db_log_dir: 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.wal_dir: /var/lib/ceph/mon/ceph-vm08/store.db 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-09T13:53:34.685 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.advise_random_on_open: 1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.db_write_buffer_size: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.write_buffer_manager: 0x55a7a09ea2a0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.new_table_reader_for_compaction_inputs: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.rate_limiter: (nil) 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.wal_recovery_mode: 2 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.enable_thread_tracking: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.enable_pipelined_write: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.unordered_write: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.row_cache: None 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.wal_filter: None 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.allow_ingest_behind: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.preserve_deletes: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.two_write_queues: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.manual_wal_flush: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.atomic_flush: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.log_readahead_size: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.best_efforts_recovery: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.allow_data_in_errors: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.db_host_id: __hostname__ 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_background_jobs: 2 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_background_compactions: -1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_subcompactions: 1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_total_wal_size: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_open_files: -1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bytes_per_sync: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_readahead_size: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_background_flushes: -1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Compression algorithms supported: 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: kZSTD supported: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: kXpressCompression supported: 0 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: kLZ4HCCompression supported: 1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: kLZ4Compression supported: 1 2026-03-09T13:53:34.686 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: kBZip2Compression supported: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: kZlibCompression supported: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: kSnappyCompression supported: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/version_set.cc:4725] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm08/store.db/MANIFEST-000009 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/column_family.cc:597] --------------- Options for column family [default]: 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.merge_operator: 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_filter: None 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_filter_factory: None 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.sst_partitioner_factory: None 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55a7a08c5d00) 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: cache_index_and_filter_blocks: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: pin_top_level_index_and_filter: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: index_type: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: data_block_index_type: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: index_shortening: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: hash_index_allow_collision: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: checksum: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: no_block_cache: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: block_cache: 0x55a7a0930170 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: block_cache_name: BinnedLRUCache 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: block_cache_options: 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: capacity : 536870912 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: num_shard_bits : 4 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: strict_capacity_limit : 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: high_pri_pool_ratio: 0.000 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: block_cache_compressed: (nil) 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: persistent_cache: (nil) 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: block_size: 4096 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: block_size_deviation: 10 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: block_restart_interval: 16 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: index_block_restart_interval: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: metadata_block_size: 4096 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: partition_filters: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: use_delta_encoding: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: filter_policy: rocksdb.BuiltinBloomFilter 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: whole_key_filtering: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: verify_compression: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: read_amp_bytes_per_bit: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: format_version: 4 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: enable_index_compression: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout: block_align: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.write_buffer_size: 33554432 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_write_buffer_number: 2 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compression: NoCompression 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bottommost_compression: Disabled 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.prefix_extractor: nullptr 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.num_levels: 7 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-09T13:53:34.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compression_opts.level: 32767 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compression_opts.strategy: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compression_opts.enabled: false 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.target_file_size_base: 67108864 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.arena_block_size: 4194304 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.rate_limit_delay_max_milliseconds: 100 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.disable_auto_compactions: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.table_properties_collectors: 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.inplace_update_support: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.bloom_locality: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.max_successive_merges: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.paranoid_file_checks: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.force_consistency_checks: 1 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.report_bg_io_stats: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.ttl: 2592000 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-09T13:53:34.688 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.enable_blob_files: false 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.min_blob_size: 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.blob_file_size: 268435456 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/version_set.cc:4773] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm08/store.db/MANIFEST-000009 succeeded,manifest_file_number is 9, next_file_number is 11, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/version_set.cc:4782] Column family [default] (ID 0), log number is 5 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/version_set.cc:4083] Creating manifest 13 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773064414483614, "job": 1, "event": "recovery_started", "wal_files": [10]} 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/db_impl/db_impl_open.cc:847] Recovering log #10 mode 2 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [table/block_based/filter_policy.cc:996] Using legacy Bloom filter with high (20) bits/key. Dramatic filter space and/or accuracy improvement is available with format_version>=5. 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773064414486920, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 14, "file_size": 77521, "file_checksum": "", "file_checksum_func_name": "Unknown", "table_properties": {"data_size": 75808, "index_size": 206, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 581, "raw_key_size": 9799, "raw_average_key_size": 48, "raw_value_size": 70201, "raw_average_value_size": 344, "num_data_blocks": 9, "num_entries": 204, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "rocksdb.BuiltinBloomFilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; ", "creation_time": 1773064414, "oldest_key_time": 0, "file_creation_time": 0, "db_id": "150466dc-42a1-4a13-a1a6-ebec8d02e511", "db_session_id": "YB7R1PJK8OJ2G281WZFI"}} 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/version_set.cc:4083] Creating manifest 15 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773064414491315, "job": 1, "event": "recovery_finished"} 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [file/delete_scheduler.cc:73] Deleted file /var/lib/ceph/mon/ceph-vm08/store.db/000010.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/db_impl/db_impl_open.cc:1701] SstFileManager instance 0x55a7a0916700 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: DB pointer 0x55a7a098a000 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/db_impl/db_impl.cc:902] ------- DUMPING STATS ------- 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: rocksdb: [db/db_impl/db_impl.cc:903] 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ** DB Stats ** 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 MB, 0.00 MB/s 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ** Compaction Stats [default] ** 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: L0 2/0 77.48 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 56.7 0.00 0.00 1 0.001 0 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Sum 2/0 77.48 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 56.7 0.00 0.00 1 0.001 0 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 56.7 0.00 0.00 1 0.001 0 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ** Compaction Stats [default] ** 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 56.7 0.00 0.00 1 0.001 0 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Cumulative compaction: 0.00 GB write, 3.05 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Interval compaction: 0.00 GB write, 3.05 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ** Compaction Stats [default] ** 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: L0 2/0 77.48 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 56.7 0.00 0.00 1 0.001 0 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Sum 2/0 77.48 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 56.7 0.00 0.00 1 0.001 0 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: 2026-03-09T13:53:34.689 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ** Compaction Stats [default] ** 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 56.7 0.00 0.00 1 0.001 0 0 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Cumulative compaction: 0.00 GB write, 3.05 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: starting mon.vm08 rank 0 at public addrs [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] at bind addrs [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon_data /var/lib/ceph/mon/ceph-vm08 fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???) e1 preinit fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).mds e0 Unable to load 'last_metadata' 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).mds e1 new map 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).mds e1 print_map 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: e1 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2} 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: legacy client fscid: -1 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout: No filesystems configured 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).mgr e0 loading version 1 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).mgr e1 active server: (0) 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08@-1(???).mgr e1 mkfs or daemon transitioned to available, loading commands 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mon.vm08 is new leader, mons vm08 in quorum (ranks 0) 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: monmap e1: 1 mons at {vm08=[v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0]} 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: fsmap 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: osdmap e1: 0 total, 0 up, 0 in 2026-03-09T13:53:34.690 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:34 vm08 ceph-mon[47734]: mgrmap e1: no daemons active 2026-03-09T13:53:34.712 INFO:teuthology.orchestra.run.vm08.stderr:Wrote config to /etc/ceph/ceph.conf 2026-03-09T13:53:34.712 INFO:teuthology.orchestra.run.vm08.stderr:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-09T13:53:34.712 INFO:teuthology.orchestra.run.vm08.stderr:Creating mgr... 2026-03-09T13:53:34.712 INFO:teuthology.orchestra.run.vm08.stderr:Verifying port 9283 ... 2026-03-09T13:53:34.910 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Failed to reset failed state of unit ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mgr.vm08.siqces.service: Unit ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mgr.vm08.siqces.service not loaded. 2026-03-09T13:53:34.926 INFO:teuthology.orchestra.run.vm08.stderr:systemctl: Created symlink /etc/systemd/system/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64.target.wants/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mgr.vm08.siqces.service → /etc/systemd/system/ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@.service. 2026-03-09T13:53:35.379 INFO:teuthology.orchestra.run.vm08.stderr:firewalld does not appear to be present 2026-03-09T13:53:35.379 INFO:teuthology.orchestra.run.vm08.stderr:Not possible to enable service . firewalld.service is not available 2026-03-09T13:53:35.379 INFO:teuthology.orchestra.run.vm08.stderr:firewalld does not appear to be present 2026-03-09T13:53:35.379 INFO:teuthology.orchestra.run.vm08.stderr:Not possible to open ports <[9283]>. firewalld.service is not available 2026-03-09T13:53:35.379 INFO:teuthology.orchestra.run.vm08.stderr:Waiting for mgr to start... 2026-03-09T13:53:35.379 INFO:teuthology.orchestra.run.vm08.stderr:Waiting for mgr... 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: { 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "fsid": "4a3e67a6-1bbf-11f1-aca1-21c5f680cc64", 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "health": { 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "checks": {}, 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mutes": [] 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum": [ 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 0 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "vm08" 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum_age": 1, 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "monmap": { 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osdmap": { 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:35.698 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "pgmap": { 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:35.699 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "fsmap": { 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "available": false, 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "modules": [ 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "iostat", 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "nfs", 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "restful" 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "services": {} 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "servicemap": { 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "modified": "2026-03-09T13:53:32.662476+0000", 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "services": {} 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-09T13:53:35.700 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: } 2026-03-09T13:53:35.745 INFO:teuthology.orchestra.run.vm08.stderr:mgr not available, waiting (1/15)... 2026-03-09T13:53:35.962 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:35 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1476670705' entity='client.admin' 2026-03-09T13:53:36.720 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:36 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/894068071' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T13:53:38.127 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: { 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "fsid": "4a3e67a6-1bbf-11f1-aca1-21c5f680cc64", 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "health": { 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "checks": {}, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mutes": [] 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum": [ 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 0 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "vm08" 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum_age": 3, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "monmap": { 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osdmap": { 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-09T13:53:38.128 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "pgmap": { 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:38.129 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "fsmap": { 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "available": false, 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "modules": [ 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "iostat", 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "nfs", 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "restful" 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "services": {} 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "servicemap": { 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "modified": "2026-03-09T13:53:32.662476+0000", 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "services": {} 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-09T13:53:38.130 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: } 2026-03-09T13:53:38.178 INFO:teuthology.orchestra.run.vm08.stderr:mgr not available, waiting (2/15)... 2026-03-09T13:53:38.247 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:38 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1632181719' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T13:53:40.491 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:40.492 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: { 2026-03-09T13:53:40.492 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "fsid": "4a3e67a6-1bbf-11f1-aca1-21c5f680cc64", 2026-03-09T13:53:40.492 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "health": { 2026-03-09T13:53:40.492 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-09T13:53:40.492 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "checks": {}, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mutes": [] 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum": [ 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 0 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "vm08" 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum_age": 5, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "monmap": { 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osdmap": { 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-09T13:53:40.493 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-09T13:53:40.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:40.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "pgmap": { 2026-03-09T13:53:40.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-09T13:53:40.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-09T13:53:40.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-09T13:53:40.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-09T13:53:40.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-09T13:53:40.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "fsmap": { 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "available": false, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "modules": [ 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "iostat", 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "nfs", 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "restful" 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "services": {} 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "servicemap": { 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "modified": "2026-03-09T13:53:32.662476+0000", 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "services": {} 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-09T13:53:40.495 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: } 2026-03-09T13:53:40.525 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:40 vm08 ceph-mon[47734]: Activating manager daemon vm08.siqces 2026-03-09T13:53:40.526 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:40 vm08 ceph-mon[47734]: mgrmap e2: vm08.siqces(active, starting, since 0.00485012s) 2026-03-09T13:53:40.526 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:40 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T13:53:40.526 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:40 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T13:53:40.526 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:40 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T13:53:40.526 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:40 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:53:40.526 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:40 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr metadata", "who": "vm08.siqces", "id": "vm08.siqces"}]: dispatch 2026-03-09T13:53:40.526 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:40 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3841947416' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T13:53:40.567 INFO:teuthology.orchestra.run.vm08.stderr:mgr not available, waiting (3/15)... 2026-03-09T13:53:41.563 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:41 vm08 ceph-mon[47734]: Manager daemon vm08.siqces is now available 2026-03-09T13:53:41.563 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:41 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:53:41.563 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:41 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:53:41.563 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:41 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' 2026-03-09T13:53:41.563 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:41 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' 2026-03-09T13:53:41.563 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:41 vm08 ceph-mon[47734]: from='mgr.14100 192.168.123.108:0/751994040' entity='mgr.vm08.siqces' 2026-03-09T13:53:42.567 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:42 vm08 ceph-mon[47734]: mgrmap e3: vm08.siqces(active, since 1.05192s) 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: { 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "fsid": "4a3e67a6-1bbf-11f1-aca1-21c5f680cc64", 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "health": { 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "checks": {}, 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mutes": [] 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum": [ 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 0 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "vm08" 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "quorum_age": 8, 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "monmap": { 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osdmap": { 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-09T13:53:42.973 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "pgmap": { 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "fsmap": { 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "available": true, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "modules": [ 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "iostat", 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "nfs", 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "restful" 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ], 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "services": {} 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "servicemap": { 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "modified": "2026-03-09T13:53:32.662476+0000", 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "services": {} 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: }, 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-09T13:53:42.974 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: } 2026-03-09T13:53:43.012 INFO:teuthology.orchestra.run.vm08.stderr:mgr is available 2026-03-09T13:53:43.270 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:43.270 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: [global] 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: fsid = 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mon_osd_allow_pg_remap = true 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mon_osd_allow_primary_affinity = true 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mon_warn_on_no_sortbitwise = false 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd_crush_chooseleaf_type = 0 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: [mgr] 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: mgr/telemetry/nag = false 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: [osd] 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd_map_max_advance = 10 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd_mclock_iops_capacity_threshold_hdd = 49000 2026-03-09T13:53:43.271 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: osd_sloppy_crc = true 2026-03-09T13:53:43.308 INFO:teuthology.orchestra.run.vm08.stderr:Enabling cephadm module... 2026-03-09T13:53:43.575 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:43 vm08 ceph-mon[47734]: mgrmap e4: vm08.siqces(active, since 2s) 2026-03-09T13:53:43.575 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:43 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2823859487' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T13:53:43.575 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:43 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1396371338' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T13:53:44.588 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:44 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/358387980' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-09T13:53:44.955 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: { 2026-03-09T13:53:44.955 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 5, 2026-03-09T13:53:44.955 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "available": true, 2026-03-09T13:53:44.955 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "active_name": "vm08.siqces", 2026-03-09T13:53:44.955 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_standby": 0 2026-03-09T13:53:44.955 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: } 2026-03-09T13:53:45.029 INFO:teuthology.orchestra.run.vm08.stderr:Waiting for the mgr to restart... 2026-03-09T13:53:45.029 INFO:teuthology.orchestra.run.vm08.stderr:Waiting for mgr epoch 5... 2026-03-09T13:53:45.605 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:45 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/358387980' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-09T13:53:45.605 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:45 vm08 ceph-mon[47734]: mgrmap e5: vm08.siqces(active, since 4s) 2026-03-09T13:53:45.605 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:45 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/359044745' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T13:53:49.644 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:49 vm08 ceph-mon[47734]: Active manager daemon vm08.siqces restarted 2026-03-09T13:53:49.644 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:49 vm08 ceph-mon[47734]: Activating manager daemon vm08.siqces 2026-03-09T13:53:49.644 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:49 vm08 ceph-mon[47734]: osdmap e2: 0 total, 0 up, 0 in 2026-03-09T13:53:50.512 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: { 2026-03-09T13:53:50.513 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mgrmap_epoch": 7, 2026-03-09T13:53:50.513 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "initialized": true 2026-03-09T13:53:50.513 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: } 2026-03-09T13:53:50.541 INFO:teuthology.orchestra.run.vm08.stderr:mgr epoch 5 is available 2026-03-09T13:53:50.541 INFO:teuthology.orchestra.run.vm08.stderr:Setting orchestrator backend to cephadm... 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: mgrmap e6: vm08.siqces(active, starting, since 0.0544235s) 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr metadata", "who": "vm08.siqces", "id": "vm08.siqces"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: Manager daemon vm08.siqces is now available 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:50.651 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:50 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:53:51.120 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: value unchanged 2026-03-09T13:53:51.161 INFO:teuthology.orchestra.run.vm08.stderr:Generating ssh key... 2026-03-09T13:53:51.671 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:51 vm08 ceph-mon[47734]: [09/Mar/2026:13:53:50] ENGINE Bus STARTING 2026-03-09T13:53:51.671 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:51 vm08 ceph-mon[47734]: [09/Mar/2026:13:53:50] ENGINE Serving on https://192.168.123.108:7150 2026-03-09T13:53:51.671 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:51 vm08 ceph-mon[47734]: [09/Mar/2026:13:53:50] ENGINE Bus STARTED 2026-03-09T13:53:51.671 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:51 vm08 ceph-mon[47734]: from='client.14124 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T13:53:51.671 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:51 vm08 ceph-mon[47734]: mgrmap e7: vm08.siqces(active, since 1.07795s) 2026-03-09T13:53:51.671 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:51 vm08 ceph-mon[47734]: from='client.14124 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T13:53:51.671 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:51 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:51.671 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:51 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:53:52.010 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDjxQ0RLPdw0b+a248YiB4qkB1I2SbqSnIboaV+tG97EV+4klgjWkpawdrFVEogJ0Af9yslmau+tHIKN3g83QuUA8kn4cIDv8/Eh5M+Jx4+KjjdcpEfd6z1JgXiZU3p7s+CTEkQz+L6psGKNMfVeiUvAu9MFHLb/B6Vk7DLwbQKS1OnYwPYU5Rfo3HGU3CYxmlDNZibk2/T5/9P1EgYH1w6440lJAhCpu4iX+gxgJSw6BtYISA1tyqGUjK4s/UIXOGV7BCXKhqzMBabra5K8pySPnJhu1rPAsn1mUXWNjV5qI25v+AiE8OI4iYRD++lxpEDOo5nu2NUYtpZgVYNokrI6ZIbBdX+41efju0QKso0GlHStJKuIOAMzbmTgtzBSejYECV8+eHMdjfh4PFtCeIXbT8T8ctxE/gqfd/kjZcJT0XJj74axfScnNGCiemQpYpvvv4x9e74PYumqxmZBaH3kKv1Jn+zugJIQ2DzlFhd/J8PDCUS4/z6dNK1mQCrsp0= ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:53:52.063 INFO:teuthology.orchestra.run.vm08.stderr:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-09T13:53:52.064 INFO:teuthology.orchestra.run.vm08.stderr:Adding key to root@localhost authorized_keys... 2026-03-09T13:53:52.064 INFO:teuthology.orchestra.run.vm08.stderr:Adding host vm08... 2026-03-09T13:53:52.589 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:52 vm08 ceph-mon[47734]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:52.589 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:52 vm08 ceph-mon[47734]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:52.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:52 vm08 ceph-mon[47734]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:52.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:52 vm08 ceph-mon[47734]: Generating ssh key... 2026-03-09T13:53:52.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:52 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:52.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:52 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:52.590 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:52 vm08 ceph-mon[47734]: mgrmap e8: vm08.siqces(active, since 2s) 2026-03-09T13:53:52.952 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Added host 'vm08' with addr '192.168.123.108' 2026-03-09T13:53:52.990 INFO:teuthology.orchestra.run.vm08.stderr:Deploying mon service with default placement... 2026-03-09T13:53:53.259 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Scheduled mon update... 2026-03-09T13:53:53.293 INFO:teuthology.orchestra.run.vm08.stderr:Deploying mgr service with default placement... 2026-03-09T13:53:53.510 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:53 vm08 ceph-mon[47734]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:53.510 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:53 vm08 ceph-mon[47734]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm08", "addr": "192.168.123.108", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:53.510 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:53 vm08 ceph-mon[47734]: Deploying cephadm binary to vm08 2026-03-09T13:53:53.510 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:53 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:53.510 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:53 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:53:53.510 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:53 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:53.556 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Scheduled mgr update... 2026-03-09T13:53:53.610 INFO:teuthology.orchestra.run.vm08.stderr:Deploying crash service with default placement... 2026-03-09T13:53:53.866 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Scheduled crash update... 2026-03-09T13:53:53.912 INFO:teuthology.orchestra.run.vm08.stderr:Deploying prometheus service with default placement... 2026-03-09T13:53:54.360 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Scheduled prometheus update... 2026-03-09T13:53:54.413 INFO:teuthology.orchestra.run.vm08.stderr:Deploying grafana service with default placement... 2026-03-09T13:53:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:54 vm08 ceph-mon[47734]: Added host vm08 2026-03-09T13:53:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:54 vm08 ceph-mon[47734]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:54 vm08 ceph-mon[47734]: Saving service mon spec with placement count:5 2026-03-09T13:53:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:54 vm08 ceph-mon[47734]: from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:54 vm08 ceph-mon[47734]: Saving service mgr spec with placement count:2 2026-03-09T13:53:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:54 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:54.620 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:54 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:54.620 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:54 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:54.685 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Scheduled grafana update... 2026-03-09T13:53:54.751 INFO:teuthology.orchestra.run.vm08.stderr:Deploying node-exporter service with default placement... 2026-03-09T13:53:55.053 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Scheduled node-exporter update... 2026-03-09T13:53:55.101 INFO:teuthology.orchestra.run.vm08.stderr:Deploying alertmanager service with default placement... 2026-03-09T13:53:55.378 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Scheduled alertmanager update... 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "crash", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: Saving service crash spec with placement * 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: Saving service prometheus spec with placement count:1 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:55.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:55 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:56.148 INFO:teuthology.orchestra.run.vm08.stderr:Enabling the dashboard module... 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: from='client.14150 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: Saving service grafana spec with placement count:1 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: from='client.14152 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: Saving service node-exporter spec with placement * 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: from='client.14154 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: Saving service alertmanager spec with placement count:1 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1308776965' entity='client.admin' 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/814630487' entity='client.admin' 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: from='mgr.14120 192.168.123.108:0/2347461357' entity='mgr.vm08.siqces' 2026-03-09T13:53:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:56 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3044828119' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-09T13:53:57.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: { 2026-03-09T13:53:57.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "epoch": 9, 2026-03-09T13:53:57.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "available": true, 2026-03-09T13:53:57.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "active_name": "vm08.siqces", 2026-03-09T13:53:57.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "num_standby": 0 2026-03-09T13:53:57.494 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: } 2026-03-09T13:53:57.523 INFO:teuthology.orchestra.run.vm08.stderr:Waiting for the mgr to restart... 2026-03-09T13:53:57.523 INFO:teuthology.orchestra.run.vm08.stderr:Waiting for mgr epoch 9... 2026-03-09T13:53:58.399 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:58 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3044828119' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-09T13:53:58.399 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:58 vm08 ceph-mon[47734]: mgrmap e9: vm08.siqces(active, since 7s) 2026-03-09T13:53:58.399 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:53:58 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2885953097' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T13:54:02.707 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: { 2026-03-09T13:54:02.709 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "mgrmap_epoch": 11, 2026-03-09T13:54:02.709 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: "initialized": true 2026-03-09T13:54:02.709 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: } 2026-03-09T13:54:02.775 INFO:teuthology.orchestra.run.vm08.stderr:mgr epoch 9 is available 2026-03-09T13:54:02.775 INFO:teuthology.orchestra.run.vm08.stderr:Generating a dashboard self-signed certificate... 2026-03-09T13:54:02.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: Active manager daemon vm08.siqces restarted 2026-03-09T13:54:02.833 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: Activating manager daemon vm08.siqces 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: osdmap e3: 0 total, 0 up, 0 in 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: mgrmap e10: vm08.siqces(active, starting, since 0.0665899s) 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr metadata", "who": "vm08.siqces", "id": "vm08.siqces"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: Manager daemon vm08.siqces is now available 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: [09/Mar/2026:13:54:02] ENGINE Bus STARTING 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:02.834 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:02 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:03.289 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: Self-signed certificate created 2026-03-09T13:54:03.322 INFO:teuthology.orchestra.run.vm08.stderr:Creating initial admin user... 2026-03-09T13:54:03.743 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: {"username": "admin", "password": "$2b$12$lhsOfpSgdTMOCt6cHgzbo.1LZn1a1/CL8M7eyYLRJyjISl2NUZ81O", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773064443, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-09T13:54:03.777 INFO:teuthology.orchestra.run.vm08.stderr:Fetching dashboard port number... 2026-03-09T13:54:04.015 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:03 vm08 ceph-mon[47734]: [09/Mar/2026:13:54:02] ENGINE Serving on https://192.168.123.108:7150 2026-03-09T13:54:04.015 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:03 vm08 ceph-mon[47734]: [09/Mar/2026:13:54:02] ENGINE Bus STARTED 2026-03-09T13:54:04.015 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:03 vm08 ceph-mon[47734]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T13:54:04.015 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:03 vm08 ceph-mon[47734]: mgrmap e11: vm08.siqces(active, since 1.0709s) 2026-03-09T13:54:04.015 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:03 vm08 ceph-mon[47734]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T13:54:04.015 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:03 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:04.015 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:03 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:04.015 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:03 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:04.016 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: 8443 2026-03-09T13:54:04.061 INFO:teuthology.orchestra.run.vm08.stderr:firewalld does not appear to be present 2026-03-09T13:54:04.061 INFO:teuthology.orchestra.run.vm08.stderr:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-09T13:54:04.062 INFO:teuthology.orchestra.run.vm08.stderr:Ceph Dashboard is now available at: 2026-03-09T13:54:04.062 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:54:04.062 INFO:teuthology.orchestra.run.vm08.stderr: URL: https://vm08.local:8443/ 2026-03-09T13:54:04.062 INFO:teuthology.orchestra.run.vm08.stderr: User: admin 2026-03-09T13:54:04.062 INFO:teuthology.orchestra.run.vm08.stderr: Password: cp3w2igv2z 2026-03-09T13:54:04.062 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:54:04.062 INFO:teuthology.orchestra.run.vm08.stderr:Enabling autotune for osd_memory_target 2026-03-09T13:54:04.649 INFO:teuthology.orchestra.run.vm08.stderr:/usr/bin/ceph: set mgr/dashboard/cluster/status 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr:You can access the Ceph CLI with: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: sudo /home/ubuntu/cephtest/cephadm shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr:Please consider enabling telemetry to help improve Ceph: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: ceph telemetry on 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr:For more information see: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: https://docs.ceph.com/docs/master/mgr/telemetry/ 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr: 2026-03-09T13:54:04.701 INFO:teuthology.orchestra.run.vm08.stderr:Bootstrap complete. 2026-03-09T13:54:04.726 INFO:tasks.cephadm:Fetching config... 2026-03-09T13:54:04.726 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:54:04.726 DEBUG:teuthology.orchestra.run.vm08:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-09T13:54:04.748 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-09T13:54:04.749 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:54:04.749 DEBUG:teuthology.orchestra.run.vm08:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-09T13:54:04.827 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-09T13:54:04.827 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:54:04.827 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/keyring of=/dev/stdout 2026-03-09T13:54:04.898 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-09T13:54:04.899 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:54:04.899 DEBUG:teuthology.orchestra.run.vm08:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-09T13:54:04.957 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-09T13:54:04.957 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDjxQ0RLPdw0b+a248YiB4qkB1I2SbqSnIboaV+tG97EV+4klgjWkpawdrFVEogJ0Af9yslmau+tHIKN3g83QuUA8kn4cIDv8/Eh5M+Jx4+KjjdcpEfd6z1JgXiZU3p7s+CTEkQz+L6psGKNMfVeiUvAu9MFHLb/B6Vk7DLwbQKS1OnYwPYU5Rfo3HGU3CYxmlDNZibk2/T5/9P1EgYH1w6440lJAhCpu4iX+gxgJSw6BtYISA1tyqGUjK4s/UIXOGV7BCXKhqzMBabra5K8pySPnJhu1rPAsn1mUXWNjV5qI25v+AiE8OI4iYRD++lxpEDOo5nu2NUYtpZgVYNokrI6ZIbBdX+41efju0QKso0GlHStJKuIOAMzbmTgtzBSejYECV8+eHMdjfh4PFtCeIXbT8T8ctxE/gqfd/kjZcJT0XJj74axfScnNGCiemQpYpvvv4x9e74PYumqxmZBaH3kKv1Jn+zugJIQ2DzlFhd/J8PDCUS4/z6dNK1mQCrsp0= ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T13:54:05.043 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:04 vm08 ceph-mon[47734]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:54:05.043 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:04 vm08 ceph-mon[47734]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:54:05.043 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:04 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:05.043 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:04 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2495172544' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-09T13:54:05.043 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:04 vm08 ceph-mon[47734]: mgrmap e12: vm08.siqces(active, since 2s) 2026-03-09T13:54:05.044 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:04 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1070770501' entity='client.admin' 2026-03-09T13:54:05.058 INFO:teuthology.orchestra.run.vm08.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDjxQ0RLPdw0b+a248YiB4qkB1I2SbqSnIboaV+tG97EV+4klgjWkpawdrFVEogJ0Af9yslmau+tHIKN3g83QuUA8kn4cIDv8/Eh5M+Jx4+KjjdcpEfd6z1JgXiZU3p7s+CTEkQz+L6psGKNMfVeiUvAu9MFHLb/B6Vk7DLwbQKS1OnYwPYU5Rfo3HGU3CYxmlDNZibk2/T5/9P1EgYH1w6440lJAhCpu4iX+gxgJSw6BtYISA1tyqGUjK4s/UIXOGV7BCXKhqzMBabra5K8pySPnJhu1rPAsn1mUXWNjV5qI25v+AiE8OI4iYRD++lxpEDOo5nu2NUYtpZgVYNokrI6ZIbBdX+41efju0QKso0GlHStJKuIOAMzbmTgtzBSejYECV8+eHMdjfh4PFtCeIXbT8T8ctxE/gqfd/kjZcJT0XJj74axfScnNGCiemQpYpvvv4x9e74PYumqxmZBaH3kKv1Jn+zugJIQ2DzlFhd/J8PDCUS4/z6dNK1mQCrsp0= ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:54:05.071 DEBUG:teuthology.orchestra.run.vm10:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDjxQ0RLPdw0b+a248YiB4qkB1I2SbqSnIboaV+tG97EV+4klgjWkpawdrFVEogJ0Af9yslmau+tHIKN3g83QuUA8kn4cIDv8/Eh5M+Jx4+KjjdcpEfd6z1JgXiZU3p7s+CTEkQz+L6psGKNMfVeiUvAu9MFHLb/B6Vk7DLwbQKS1OnYwPYU5Rfo3HGU3CYxmlDNZibk2/T5/9P1EgYH1w6440lJAhCpu4iX+gxgJSw6BtYISA1tyqGUjK4s/UIXOGV7BCXKhqzMBabra5K8pySPnJhu1rPAsn1mUXWNjV5qI25v+AiE8OI4iYRD++lxpEDOo5nu2NUYtpZgVYNokrI6ZIbBdX+41efju0QKso0GlHStJKuIOAMzbmTgtzBSejYECV8+eHMdjfh4PFtCeIXbT8T8ctxE/gqfd/kjZcJT0XJj74axfScnNGCiemQpYpvvv4x9e74PYumqxmZBaH3kKv1Jn+zugJIQ2DzlFhd/J8PDCUS4/z6dNK1mQCrsp0= ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T13:54:05.111 INFO:teuthology.orchestra.run.vm10.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDjxQ0RLPdw0b+a248YiB4qkB1I2SbqSnIboaV+tG97EV+4klgjWkpawdrFVEogJ0Af9yslmau+tHIKN3g83QuUA8kn4cIDv8/Eh5M+Jx4+KjjdcpEfd6z1JgXiZU3p7s+CTEkQz+L6psGKNMfVeiUvAu9MFHLb/B6Vk7DLwbQKS1OnYwPYU5Rfo3HGU3CYxmlDNZibk2/T5/9P1EgYH1w6440lJAhCpu4iX+gxgJSw6BtYISA1tyqGUjK4s/UIXOGV7BCXKhqzMBabra5K8pySPnJhu1rPAsn1mUXWNjV5qI25v+AiE8OI4iYRD++lxpEDOo5nu2NUYtpZgVYNokrI6ZIbBdX+41efju0QKso0GlHStJKuIOAMzbmTgtzBSejYECV8+eHMdjfh4PFtCeIXbT8T8ctxE/gqfd/kjZcJT0XJj74axfScnNGCiemQpYpvvv4x9e74PYumqxmZBaH3kKv1Jn+zugJIQ2DzlFhd/J8PDCUS4/z6dNK1mQCrsp0= ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:54:05.122 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-09T13:54:05.768 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-09T13:54:05.768 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-09T13:54:06.369 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm10 2026-03-09T13:54:06.369 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T13:54:06.369 DEBUG:teuthology.orchestra.run.vm10:> dd of=/etc/ceph/ceph.conf 2026-03-09T13:54:06.395 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T13:54:06.395 DEBUG:teuthology.orchestra.run.vm10:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T13:54:06.453 INFO:tasks.cephadm:Adding host vm10 to orchestrator... 2026-03-09T13:54:06.453 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch host add vm10 2026-03-09T13:54:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:06 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2303862177' entity='client.admin' 2026-03-09T13:54:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:06 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:06 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:06 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:54:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:06 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:07.702 INFO:teuthology.orchestra.run.vm08.stdout:Added host 'vm10' with addr '192.168.123.110' 2026-03-09T13:54:07.751 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch host ls --format=json 2026-03-09T13:54:07.943 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:07 vm08 ceph-mon[47734]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:54:07.943 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:07 vm08 ceph-mon[47734]: Deploying daemon alertmanager.vm08 on vm08 2026-03-09T13:54:07.943 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:07 vm08 ceph-mon[47734]: from='client.14188 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm10", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:54:08.258 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:54:08.258 INFO:teuthology.orchestra.run.vm08.stdout:[{"addr": "192.168.123.108", "hostname": "vm08", "labels": [], "status": ""}, {"addr": "192.168.123.110", "hostname": "vm10", "labels": [], "status": ""}] 2026-03-09T13:54:08.463 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-09T13:54:08.463 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd crush tunables default 2026-03-09T13:54:08.908 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:08 vm08 ceph-mon[47734]: Deploying cephadm binary to vm10 2026-03-09T13:54:08.908 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:08 vm08 ceph-mon[47734]: mgrmap e13: vm08.siqces(active, since 6s) 2026-03-09T13:54:08.908 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:08 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:08.908 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:08 vm08 ceph-mon[47734]: Added host vm10 2026-03-09T13:54:09.713 INFO:teuthology.orchestra.run.vm08.stderr:adjusted tunables profile to default 2026-03-09T13:54:09.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:09 vm08 ceph-mon[47734]: from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:54:09.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:09 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1131173443' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-09T13:54:09.902 INFO:tasks.cephadm:Adding mon.vm08 on vm08 2026-03-09T13:54:09.902 INFO:tasks.cephadm:Adding mon.vm10 on vm10 2026-03-09T13:54:09.902 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch apply mon '2;vm08:192.168.123.108=vm08;vm10:192.168.123.110=vm10' 2026-03-09T13:54:10.439 INFO:teuthology.orchestra.run.vm10.stdout:Scheduled mon update... 2026-03-09T13:54:10.509 DEBUG:teuthology.orchestra.run.vm10:mon.vm10> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@mon.vm10.service 2026-03-09T13:54:10.510 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:10.510 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:10.754 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:10 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1131173443' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-09T13:54:10.754 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:10 vm08 ceph-mon[47734]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T13:54:10.754 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:10 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:10.754 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:10 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T13:54:10.754 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:10 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-09T13:54:10.754 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:10 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:54:10.754 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:10 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:11.068 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:11.068 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:11.071 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:12.141 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:12.141 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: Deploying daemon crash.vm08 on vm08 2026-03-09T13:54:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='client.14194 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm08:192.168.123.108=vm08;vm10:192.168.123.110=vm10", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:54:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: Saving service mon spec with placement vm08:192.168.123.108=vm08;vm10:192.168.123.110=vm10;count:2 2026-03-09T13:54:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-09T13:54:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-09T13:54:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: Deploying daemon grafana.vm08 on vm08 2026-03-09T13:54:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1551856440' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:12 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:12.656 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:12.656 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:12.659 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:13 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/2932050611' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:13.700 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:13.701 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:14.199 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:14.199 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:14.201 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:14 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/4282129146' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:15.240 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:15.240 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:15.837 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:15.837 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:15.840 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:16.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:15 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1741747104' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:16.906 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:16.906 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:17.426 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:17.426 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:17.429 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:17 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/3068454502' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:18.492 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:18.492 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:19.078 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:19.078 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:19.082 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:19 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/155326959' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:20.151 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:20.151 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:20.645 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:20.645 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:20.647 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:20 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1056098754' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:21.693 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:21.693 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:22.197 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:22.197 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:22.199 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:22 vm08 ceph-mon[47734]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:22 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/3016658672' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:23.238 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:23.238 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:23.740 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:23.740 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:23.743 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:24.801 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:24.802 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:24 vm08 ceph-mon[47734]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:24 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/3815851238' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:25.310 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:25.310 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:25.312 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:26.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:25 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/2697668347' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:26.378 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:26.378 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:26.875 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:26.875 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:26.879 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:27.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:26 vm08 ceph-mon[47734]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:27.965 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:27.965 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:28.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:27 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/3840156192' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:28.494 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:28.494 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:28.496 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:28 vm08 ceph-mon[47734]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:29.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:28 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/254550386' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:29.541 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:29.541 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:30.077 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:30.078 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:30.081 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:30 vm08 ceph-mon[47734]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:30 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1609089714' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:31.136 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:31.136 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:31.638 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:31.639 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:31.641 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:32.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:31 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/536006248' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:32.710 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:32.710 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:33.234 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:33.234 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:33.236 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:33.311 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:32 vm08 ceph-mon[47734]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:34.236 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:33 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1603344809' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:34.307 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:34.308 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:34.825 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:34.825 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:34.827 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:34 vm08 ceph-mon[47734]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:34 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:35.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:34 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/742263869' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:35.894 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:35.894 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:36.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:35 vm08 ceph-mon[47734]: Deploying daemon node-exporter.vm08 on vm08 2026-03-09T13:54:36.454 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:36.454 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:36.456 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:36 vm08 ceph-mon[47734]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:36 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/2853319563' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:36 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:37.502 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:37.502 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:37.970 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:37.970 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:37.972 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:38 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' 2026-03-09T13:54:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:38 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-09T13:54:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:38 vm08 ceph-mon[47734]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:54:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:38 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/590283185' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:39.039 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:39.040 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:39.811 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:39.811 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:39.814 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:40.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:39 vm08 ceph-mon[47734]: from='mgr.14162 192.168.123.108:0/2093728261' entity='mgr.vm08.siqces' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-09T13:54:40.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:39 vm08 ceph-mon[47734]: mgrmap e14: vm08.siqces(active, since 36s) 2026-03-09T13:54:40.880 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:40.880 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:40 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/702718797' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:41.428 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:41.428 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:41.431 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:42.185 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:41 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1610051707' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:42.494 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:42.494 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:43.029 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:42 vm08 ceph-mon[47734]: Active manager daemon vm08.siqces restarted 2026-03-09T13:54:43.029 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:42 vm08 ceph-mon[47734]: Activating manager daemon vm08.siqces 2026-03-09T13:54:43.029 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:42 vm08 ceph-mon[47734]: osdmap e5: 0 total, 0 up, 0 in 2026-03-09T13:54:43.032 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:43.032 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:43.035 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: mgrmap e15: vm08.siqces(active, starting, since 0.0557045s) 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr metadata", "who": "vm08.siqces", "id": "vm08.siqces"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: Manager daemon vm08.siqces is now available 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1649879992' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: [09/Mar/2026:13:54:43] ENGINE Bus STARTING 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: [09/Mar/2026:13:54:43] ENGINE Serving on https://192.168.123.108:7150 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: [09/Mar/2026:13:54:43] ENGINE Bus STARTED 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:44.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:44.038 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:44.087 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:44.087 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:44.743 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:44.743 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:44.748 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:44 vm08 ceph-mon[47734]: mgrmap e16: vm08.siqces(active, since 1.06314s) 2026-03-09T13:54:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:54:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:44 vm08 ceph-mon[47734]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T13:54:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:44 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/2507040638' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:45.802 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:45.802 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: Updating vm08:/etc/ceph/ceph.client.admin.keyring 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm10", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm10", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:54:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:45 vm08 ceph-mon[47734]: mgrmap e17: vm08.siqces(active, since 2s) 2026-03-09T13:54:47.003 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:47.003 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:47.008 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:46 vm08 ceph-mon[47734]: Deploying daemon crash.vm10 on vm10 2026-03-09T13:54:48.061 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:48.061 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:48.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:48.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm10.euefis", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T13:54:48.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm10.euefis", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-09T13:54:48.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T13:54:48.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:54:48.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: Deploying daemon mgr.vm10.euefis on vm10 2026-03-09T13:54:48.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/2053458243' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:48.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:48.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T13:54:48.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:54:48.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:49.157 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:49.157 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":1,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:53:31.291498Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T13:54:49.169 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 1 2026-03-09T13:54:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:49 vm08 ceph-mon[47734]: Deploying daemon mon.vm10 on vm10 2026-03-09T13:54:50.230 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T13:54:50.230 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mon dump -f json 2026-03-09T13:54:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:54:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: mon.vm08 calling monitor election 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: mon.vm10 calling monitor election 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.? 192.168.123.110:0/3292074357' entity='mgr.vm10.euefis' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm10.euefis/crt"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: mon.vm08 is new leader, mons vm08,vm10 in quorum (ranks 0,1) 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: monmap e2: 2 mons at {vm08=[v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0],vm10=[v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0]} 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: fsmap 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: osdmap e5: 0 total, 0 up, 0 in 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: mgrmap e17: vm08.siqces(active, since 11s) 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: Standby manager daemon vm10.euefis started 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.? 192.168.123.110:0/3292074357' entity='mgr.vm10.euefis' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.? 192.168.123.110:0/3292074357' entity='mgr.vm10.euefis' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm10.euefis/key"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.? 192.168.123.110:0/3292074357' entity='mgr.vm10.euefis' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: overall HEALTH_OK 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:55.160 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:54:54 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:54:55.205 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:54:55.205 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":2,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","modified":"2026-03-09T13:54:49.590079Z","created":"2026-03-09T13:53:31.291498Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm08","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:3300","nonce":0},{"type":"v1","addr":"192.168.123.108:6789","nonce":0}]},"addr":"192.168.123.108:6789/0","public_addr":"192.168.123.108:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-09T13:54:55.207 INFO:teuthology.orchestra.run.vm10.stderr:dumped monmap epoch 2 2026-03-09T13:54:55.254 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-09T13:54:55.254 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph config generate-minimal-conf 2026-03-09T13:54:55.702 INFO:teuthology.orchestra.run.vm08.stdout:# minimal ceph.conf for 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:54:55.702 INFO:teuthology.orchestra.run.vm08.stdout:[global] 2026-03-09T13:54:55.702 INFO:teuthology.orchestra.run.vm08.stdout: fsid = 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 2026-03-09T13:54:55.702 INFO:teuthology.orchestra.run.vm08.stdout: mon_host = [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] 2026-03-09T13:54:55.764 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-09T13:54:55.764 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:54:55.764 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T13:54:55.787 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:55 vm08 ceph-mon[47734]: Deploying daemon prometheus.vm08 on vm08 2026-03-09T13:54:55.787 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:55 vm08 ceph-mon[47734]: mgrmap e18: vm08.siqces(active, since 11s), standbys: vm10.euefis 2026-03-09T13:54:55.787 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:55 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr metadata", "who": "vm10.euefis", "id": "vm10.euefis"}]: dispatch 2026-03-09T13:54:55.787 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:55 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1662511388' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:55.787 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:55 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:54:55.790 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:54:55.790 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T13:54:55.855 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T13:54:55.855 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T13:54:55.882 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T13:54:55.882 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T13:54:55.946 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:54:55 vm10 ceph-mon[51861]: Deploying daemon prometheus.vm08 on vm08 2026-03-09T13:54:55.946 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:54:55 vm10 ceph-mon[51861]: mgrmap e18: vm08.siqces(active, since 11s), standbys: vm10.euefis 2026-03-09T13:54:55.946 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:54:55 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr metadata", "who": "vm10.euefis", "id": "vm10.euefis"}]: dispatch 2026-03-09T13:54:55.946 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:54:55 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/1662511388' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T13:54:55.946 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:54:55 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:54:55.950 INFO:tasks.cephadm:Deploying OSDs... 2026-03-09T13:54:55.950 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:54:55.950 DEBUG:teuthology.orchestra.run.vm08:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T13:54:55.966 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T13:54:55.966 DEBUG:teuthology.orchestra.run.vm08:> ls /dev/[sv]d? 2026-03-09T13:54:56.025 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vda 2026-03-09T13:54:56.025 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdb 2026-03-09T13:54:56.025 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdc 2026-03-09T13:54:56.025 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vdd 2026-03-09T13:54:56.025 INFO:teuthology.orchestra.run.vm08.stdout:/dev/vde 2026-03-09T13:54:56.025 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T13:54:56.025 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T13:54:56.026 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdb 2026-03-09T13:54:56.085 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdb 2026-03-09T13:54:56.086 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T13:54:56.086 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-09T13:54:56.086 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T13:54:56.086 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T13:54:56.086 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 13:54:06.145628550 +0000 2026-03-09T13:54:56.086 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 13:54:05.848628249 +0000 2026-03-09T13:54:56.086 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 13:54:05.848628249 +0000 2026-03-09T13:54:56.086 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 13:49:32.239000000 +0000 2026-03-09T13:54:56.086 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T13:54:56.151 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T13:54:56.151 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T13:54:56.151 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000257302 s, 2.0 MB/s 2026-03-09T13:54:56.153 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T13:54:56.218 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdc 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdc 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 222 Links: 1 Device type: fc,20 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 13:54:06.206628611 +0000 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 13:54:05.850628251 +0000 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 13:54:05.850628251 +0000 2026-03-09T13:54:56.279 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 13:49:32.245000000 +0000 2026-03-09T13:54:56.279 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T13:54:56.348 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T13:54:56.349 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T13:54:56.349 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000160131 s, 3.2 MB/s 2026-03-09T13:54:56.350 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T13:54:56.412 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vdd 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vdd 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 225 Links: 1 Device type: fc,30 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 13:54:06.278628684 +0000 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 13:54:05.851628252 +0000 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 13:54:05.851628252 +0000 2026-03-09T13:54:56.473 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 13:49:32.255000000 +0000 2026-03-09T13:54:56.473 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T13:54:56.541 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T13:54:56.542 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T13:54:56.542 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000127358 s, 4.0 MB/s 2026-03-09T13:54:56.543 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T13:54:56.604 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vde 2026-03-09T13:54:56.668 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vde 2026-03-09T13:54:56.669 INFO:teuthology.orchestra.run.vm08.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T13:54:56.669 INFO:teuthology.orchestra.run.vm08.stdout:Device: 6h/6d Inode: 226 Links: 1 Device type: fc,40 2026-03-09T13:54:56.669 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T13:54:56.669 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T13:54:56.669 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-03-09 13:54:06.338628745 +0000 2026-03-09T13:54:56.669 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-03-09 13:54:05.854628256 +0000 2026-03-09T13:54:56.669 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-03-09 13:54:05.854628256 +0000 2026-03-09T13:54:56.669 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-03-09 13:49:32.263000000 +0000 2026-03-09T13:54:56.669 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T13:54:56.797 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-03-09T13:54:56.797 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-03-09T13:54:56.797 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000379901 s, 1.3 MB/s 2026-03-09T13:54:56.798 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:54:56 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1297516571' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:54:56.799 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T13:54:56.943 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T13:54:56.943 DEBUG:teuthology.orchestra.run.vm10:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T13:54:56.958 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T13:54:56.959 DEBUG:teuthology.orchestra.run.vm10:> ls /dev/[sv]d? 2026-03-09T13:54:57.016 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vda 2026-03-09T13:54:57.017 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdb 2026-03-09T13:54:57.017 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdc 2026-03-09T13:54:57.017 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdd 2026-03-09T13:54:57.017 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vde 2026-03-09T13:54:57.017 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T13:54:57.017 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T13:54:57.017 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdb 2026-03-09T13:54:57.075 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdb 2026-03-09T13:54:57.076 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T13:54:57.076 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,10 2026-03-09T13:54:57.076 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T13:54:57.076 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T13:54:57.076 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 13:54:45.281083131 +0000 2026-03-09T13:54:57.076 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 13:54:44.999085130 +0000 2026-03-09T13:54:57.076 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 13:54:44.999085130 +0000 2026-03-09T13:54:57.076 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 13:50:20.241000000 +0000 2026-03-09T13:54:57.076 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T13:54:57.138 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:54:56 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/1297516571' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:54:57.140 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T13:54:57.140 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T13:54:57.140 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000151242 s, 3.4 MB/s 2026-03-09T13:54:57.141 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T13:54:57.200 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdc 2026-03-09T13:54:57.260 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdc 2026-03-09T13:54:57.260 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T13:54:57.260 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T13:54:57.260 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T13:54:57.260 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T13:54:57.261 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 13:54:45.346082671 +0000 2026-03-09T13:54:57.261 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 13:54:45.007085073 +0000 2026-03-09T13:54:57.261 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 13:54:45.007085073 +0000 2026-03-09T13:54:57.261 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 13:50:20.254000000 +0000 2026-03-09T13:54:57.261 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T13:54:57.327 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T13:54:57.327 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T13:54:57.327 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000193791 s, 2.6 MB/s 2026-03-09T13:54:57.328 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T13:54:57.390 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdd 2026-03-09T13:54:57.448 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdd 2026-03-09T13:54:57.448 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T13:54:57.449 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T13:54:57.449 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T13:54:57.449 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T13:54:57.449 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 13:54:45.421082139 +0000 2026-03-09T13:54:57.449 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 13:54:45.001085115 +0000 2026-03-09T13:54:57.449 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 13:54:45.001085115 +0000 2026-03-09T13:54:57.449 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 13:50:20.259000000 +0000 2026-03-09T13:54:57.449 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T13:54:57.513 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T13:54:57.513 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T13:54:57.513 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000152153 s, 3.4 MB/s 2026-03-09T13:54:57.514 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T13:54:57.570 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vde 2026-03-09T13:54:57.627 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vde 2026-03-09T13:54:57.627 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T13:54:57.628 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T13:54:57.628 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T13:54:57.628 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T13:54:57.628 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 13:54:45.487081671 +0000 2026-03-09T13:54:57.628 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 13:54:45.004085094 +0000 2026-03-09T13:54:57.628 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 13:54:45.004085094 +0000 2026-03-09T13:54:57.628 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 13:50:20.271000000 +0000 2026-03-09T13:54:57.628 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T13:54:57.696 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T13:54:57.696 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T13:54:57.696 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000143498 s, 3.6 MB/s 2026-03-09T13:54:57.697 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T13:54:57.756 INFO:tasks.cephadm:Deploying osd.0 on vm08 with /dev/vde... 2026-03-09T13:54:57.756 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- lvm zap /dev/vde 2026-03-09T13:55:00.326 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:55:00.343 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch daemon add osd vm08:/dev/vde 2026-03-09T13:55:01.065 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:01.065 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:01.065 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:01.065 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:01.065 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:01.065 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:00 vm08 ceph-mon[47734]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T13:55:01.160 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:01.160 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:01.160 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:01.160 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:01.160 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:01.160 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:00 vm10 ceph-mon[51861]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: Updating vm10:/etc/ceph/ceph.client.admin.keyring 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: Reconfiguring mon.vm08 (unknown last config time)... 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: Reconfiguring daemon mon.vm08 on vm08 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2900766098' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8426b89c-8bdd-425e-8381-20b91a25bd3d"}]: dispatch 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2900766098' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8426b89c-8bdd-425e-8381-20b91a25bd3d"}]': finished 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T13:55:01.937 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: Updating vm10:/etc/ceph/ceph.client.admin.keyring 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='client.24101 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: Updating vm08:/etc/ceph/ceph.conf 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: Reconfiguring mon.vm08 (unknown last config time)... 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: Reconfiguring daemon mon.vm08 on vm08 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2900766098' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8426b89c-8bdd-425e-8381-20b91a25bd3d"}]: dispatch 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2900766098' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8426b89c-8bdd-425e-8381-20b91a25bd3d"}]': finished 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T13:55:02.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:02 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3835831671' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:03.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:02 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3835831671' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:04.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:03 vm08 ceph-mon[47734]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:04.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:03 vm10 ceph-mon[51861]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:06.180 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: Reconfiguring mgr.vm08.siqces (unknown last config time)... 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.siqces", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: Reconfiguring daemon mgr.vm08.siqces on vm08 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:06.181 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:05 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: Reconfiguring mgr.vm08.siqces (unknown last config time)... 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm08.siqces", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: Reconfiguring daemon mgr.vm08.siqces on vm08 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:06.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:05 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:06 vm08 ceph-mon[47734]: Deploying daemon osd.0 on vm08 2026-03-09T13:55:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:06 vm08 ceph-mon[47734]: Reconfiguring alertmanager.vm08 (dependencies changed)... 2026-03-09T13:55:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:06 vm08 ceph-mon[47734]: Reconfiguring daemon alertmanager.vm08 on vm08 2026-03-09T13:55:07.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:06 vm10 ceph-mon[51861]: Deploying daemon osd.0 on vm08 2026-03-09T13:55:07.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:06 vm10 ceph-mon[51861]: Reconfiguring alertmanager.vm08 (dependencies changed)... 2026-03-09T13:55:07.410 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:06 vm10 ceph-mon[51861]: Reconfiguring daemon alertmanager.vm08 on vm08 2026-03-09T13:55:08.523 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:08 vm08 ceph-mon[47734]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:08.524 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:08 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:08.635 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 0 on host 'vm08' 2026-03-09T13:55:08.710 DEBUG:teuthology.orchestra.run.vm08:osd.0> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@osd.0.service 2026-03-09T13:55:08.712 INFO:tasks.cephadm:Deploying osd.1 on vm08 with /dev/vdd... 2026-03-09T13:55:08.712 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- lvm zap /dev/vdd 2026-03-09T13:55:08.776 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:08 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:08.776 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:08 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T13:55:08.776 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:08 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:08.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:08 vm10 ceph-mon[51861]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:08.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:08 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:08.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:08 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:08.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:08 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm08", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T13:55:08.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:08 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:09 vm08 ceph-mon[47734]: Reconfiguring crash.vm08 (monmap changed)... 2026-03-09T13:55:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:09 vm08 ceph-mon[47734]: Reconfiguring daemon crash.vm08 on vm08 2026-03-09T13:55:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:09 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:09 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:09 vm08 ceph-mon[47734]: from='osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T13:55:09.557 INFO:journalctl@ceph.osd.0.vm08.stdout:Mar 09 13:55:09 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-0[57340]: 2026-03-09T13:55:09.360+0000 7f4f24da43c0 -1 osd.0 0 log_to_monitors true 2026-03-09T13:55:09.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:09 vm10 ceph-mon[51861]: Reconfiguring crash.vm08 (monmap changed)... 2026-03-09T13:55:09.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:09 vm10 ceph-mon[51861]: Reconfiguring daemon crash.vm08 on vm08 2026-03-09T13:55:09.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:09 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:09.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:09 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:09.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:09 vm10 ceph-mon[51861]: from='osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T13:55:10.126 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:55:10.141 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch daemon add osd vm08:/dev/vdd 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: Reconfiguring grafana.vm08 (dependencies changed)... 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: Reconfiguring daemon grafana.vm08 on vm08 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm10", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm10.euefis", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T13:55:10.716 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: Reconfiguring grafana.vm08 (dependencies changed)... 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: Reconfiguring daemon grafana.vm08 on vm08 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm10", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm10.euefis", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T13:55:10.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:10.968 INFO:journalctl@ceph.osd.0.vm08.stdout:Mar 09 13:55:10 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-0[57340]: 2026-03-09T13:55:10.846+0000 7f4f1b7a7700 -1 osd.0 0 waiting for initial osdmap 2026-03-09T13:55:10.969 INFO:journalctl@ceph.osd.0.vm08.stdout:Mar 09 13:55:10 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-0[57340]: 2026-03-09T13:55:10.861+0000 7f4f17941700 -1 osd.0 8 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: Reconfiguring crash.vm10 (monmap changed)... 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: Reconfiguring daemon crash.vm10 on vm10 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: Reconfiguring mgr.vm10.euefis (monmap changed)... 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: Reconfiguring daemon mgr.vm10.euefis on vm10 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: Reconfiguring mon.vm10 (monmap changed)... 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: Reconfiguring daemon mon.vm10 on vm10 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.108:9093"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.108:3000"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: osdmap e8: 1 total, 0 up, 1 in 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.108:9095"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:11.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: Reconfiguring crash.vm10 (monmap changed)... 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: Reconfiguring daemon crash.vm10 on vm10 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: Reconfiguring mgr.vm10.euefis (monmap changed)... 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: Reconfiguring daemon mgr.vm10.euefis on vm10 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: Reconfiguring mon.vm10 (monmap changed)... 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: Reconfiguring daemon mon.vm10 on vm10 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-09T13:55:11.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.108:9093"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.108:3000"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: osdmap e8: 1 total, 0 up, 1 in 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.108:9095"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:11.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: purged_snaps scrub starts 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: purged_snaps scrub ok 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.108:9093"}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.108:3000"}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.108:9095"}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='client.14280 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2129452382' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7759a684-08a6-4323-b3cb-092ae4decfe1"}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321] boot 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2129452382' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7759a684-08a6-4323-b3cb-092ae4decfe1"}]': finished 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:12.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:12 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3107202190' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: purged_snaps scrub starts 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: purged_snaps scrub ok 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.108:9093"}]: dispatch 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.108:3000"}]: dispatch 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.108:9095"}]: dispatch 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='client.14280 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T13:55:12.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2129452382' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "7759a684-08a6-4323-b3cb-092ae4decfe1"}]: dispatch 2026-03-09T13:55:12.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: osd.0 [v2:192.168.123.108:6802/2288390321,v1:192.168.123.108:6803/2288390321] boot 2026-03-09T13:55:12.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2129452382' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7759a684-08a6-4323-b3cb-092ae4decfe1"}]': finished 2026-03-09T13:55:12.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T13:55:12.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T13:55:12.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:12.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:12.911 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:12 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3107202190' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:14.531 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:14 vm08 ceph-mon[47734]: pgmap v12: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:14.531 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:14 vm08 ceph-mon[47734]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T13:55:14.531 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:14 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:14.660 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:14 vm10 ceph-mon[51861]: pgmap v12: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:14.660 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:14 vm10 ceph-mon[51861]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T13:55:14.660 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:14 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:16.615 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:16 vm08 ceph-mon[47734]: pgmap v14: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:16.615 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:16 vm08 ceph-mon[47734]: Detected new or changed devices on vm08 2026-03-09T13:55:16.615 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:16 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:16.615 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:16 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:16.615 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:16 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:16.615 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:16 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T13:55:16.615 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:16 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:16.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:16 vm10 ceph-mon[51861]: pgmap v14: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:16.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:16 vm10 ceph-mon[51861]: Detected new or changed devices on vm08 2026-03-09T13:55:16.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:16 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:16.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:16 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:16.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:16 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:16.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:16 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T13:55:16.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:16 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:17.490 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:17 vm08 ceph-mon[47734]: Deploying daemon osd.1 on vm08 2026-03-09T13:55:17.743 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:17 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:17.743 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:17 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:17.743 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:17 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:17.743 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:17 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:17.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:17 vm10 ceph-mon[51861]: Deploying daemon osd.1 on vm08 2026-03-09T13:55:17.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:17 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:17.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:17 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:17.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:17 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:17.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:17 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:18.343 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 1 on host 'vm08' 2026-03-09T13:55:18.396 DEBUG:teuthology.orchestra.run.vm08:osd.1> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@osd.1.service 2026-03-09T13:55:18.398 INFO:tasks.cephadm:Deploying osd.2 on vm08 with /dev/vdc... 2026-03-09T13:55:18.398 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- lvm zap /dev/vdc 2026-03-09T13:55:18.639 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:18 vm08 ceph-mon[47734]: pgmap v15: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:18.639 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:18 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:18.639 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:18 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:18.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:18 vm10 ceph-mon[51861]: pgmap v15: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:18.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:18 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:18.910 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:18 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:19.171 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:55:19.185 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch daemon add osd vm08:/dev/vdc 2026-03-09T13:55:19.512 INFO:journalctl@ceph.osd.1.vm08.stdout:Mar 09 13:55:19 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-1[60716]: 2026-03-09T13:55:19.321+0000 7f08e861f3c0 -1 osd.1 0 log_to_monitors true 2026-03-09T13:55:20.097 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:19 vm08 ceph-mon[47734]: pgmap v16: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:20.097 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:19 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:20.097 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:19 vm08 ceph-mon[47734]: from='osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T13:55:20.097 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:19 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:20.097 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:19 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:20.097 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:19 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:20.245 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:19 vm10 ceph-mon[51861]: pgmap v16: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:20.246 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:19 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:20.246 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:19 vm10 ceph-mon[51861]: from='osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T13:55:20.246 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:19 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:20.246 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:19 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:20.246 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:19 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:20.993 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='client.14298 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:20.993 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T13:55:20.993 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T13:55:20.993 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:20.993 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T13:55:20.993 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3419717845' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8a88652c-970b-4ebe-8175-ea8b9c059735"}]: dispatch 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3419717845' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8a88652c-970b-4ebe-8175-ea8b9c059735"}]': finished 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: osdmap e12: 3 total, 1 up, 3 in 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:20.994 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:20 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:20.994 INFO:journalctl@ceph.osd.1.vm08.stdout:Mar 09 13:55:20 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-1[60716]: 2026-03-09T13:55:20.766+0000 7f08df022700 -1 osd.1 0 waiting for initial osdmap 2026-03-09T13:55:20.994 INFO:journalctl@ceph.osd.1.vm08.stdout:Mar 09 13:55:20 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-1[60716]: 2026-03-09T13:55:20.772+0000 7f08d91b8700 -1 osd.1 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='client.14298 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3419717845' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "8a88652c-970b-4ebe-8175-ea8b9c059735"}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3419717845' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8a88652c-970b-4ebe-8175-ea8b9c059735"}]': finished 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: osdmap e12: 3 total, 1 up, 3 in 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:21.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:20 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:22.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:21 vm08 ceph-mon[47734]: pgmap v19: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:21 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1649702622' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:21 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:21 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:21 vm08 ceph-mon[47734]: osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833] boot 2026-03-09T13:55:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:21 vm08 ceph-mon[47734]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T13:55:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:21 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:21 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:21 vm10 ceph-mon[51861]: pgmap v19: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T13:55:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:21 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/1649702622' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:21 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:21 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:21 vm10 ceph-mon[51861]: osd.1 [v2:192.168.123.108:6810/3586975833,v1:192.168.123.108:6811/3586975833] boot 2026-03-09T13:55:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:21 vm10 ceph-mon[51861]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T13:55:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:21 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T13:55:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:21 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:22 vm08 ceph-mon[47734]: purged_snaps scrub starts 2026-03-09T13:55:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:22 vm08 ceph-mon[47734]: purged_snaps scrub ok 2026-03-09T13:55:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:22 vm10 ceph-mon[51861]: purged_snaps scrub starts 2026-03-09T13:55:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:22 vm10 ceph-mon[51861]: purged_snaps scrub ok 2026-03-09T13:55:24.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:24 vm08 ceph-mon[47734]: pgmap v21: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T13:55:24.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:24 vm08 ceph-mon[47734]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T13:55:24.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:24 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:24 vm10 ceph-mon[51861]: pgmap v21: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T13:55:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:24 vm10 ceph-mon[51861]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T13:55:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:24 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:25.047 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:25 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T13:55:25.047 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:25 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:25.047 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:25 vm08 ceph-mon[47734]: Deploying daemon osd.2 on vm08 2026-03-09T13:55:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:25 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T13:55:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:25 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:25 vm10 ceph-mon[51861]: Deploying daemon osd.2 on vm08 2026-03-09T13:55:26.175 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:26 vm08 ceph-mon[47734]: pgmap v23: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T13:55:26.175 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:26 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:26.175 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:26 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:26.175 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:26 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:26.175 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:26 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:26 vm10 ceph-mon[51861]: pgmap v23: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T13:55:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:26 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:26 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:26 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:26 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:26.735 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 2 on host 'vm08' 2026-03-09T13:55:26.809 DEBUG:teuthology.orchestra.run.vm08:osd.2> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@osd.2.service 2026-03-09T13:55:26.810 INFO:tasks.cephadm:Deploying osd.3 on vm08 with /dev/vdb... 2026-03-09T13:55:26.810 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- lvm zap /dev/vdb 2026-03-09T13:55:27.346 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 13:55:27 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-2[64038]: 2026-03-09T13:55:27.193+0000 7fecb8b6d3c0 -1 osd.2 0 log_to_monitors true 2026-03-09T13:55:27.613 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:55:27.630 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch daemon add osd vm08:/dev/vdb 2026-03-09T13:55:27.851 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:27 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:27.852 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:27 vm08 ceph-mon[47734]: pgmap v24: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T13:55:27.852 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:27 vm08 ceph-mon[47734]: from='osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T13:55:27.852 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:27 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:28.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:27 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:28.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:27 vm10 ceph-mon[51861]: pgmap v24: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T13:55:28.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:27 vm10 ceph-mon[51861]: from='osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T13:55:28.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:27 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:28.750 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 13:55:28 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-2[64038]: 2026-03-09T13:55:28.747+0000 7fecaf570700 -1 osd.2 0 waiting for initial osdmap 2026-03-09T13:55:29.058 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T13:55:29.058 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T13:55:29.058 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:29.058 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T13:55:29.058 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='client.14316 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: Detected new or changed devices on vm08 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:29.059 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:28 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:29.059 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 13:55:28 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-2[64038]: 2026-03-09T13:55:28.764+0000 7fecabf0b700 -1 osd.2 16 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T13:55:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T13:55:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T13:55:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T13:55:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='client.14316 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm08:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:29.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:29.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: Detected new or changed devices on vm08 2026-03-09T13:55:29.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:29.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:29.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:29.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:29.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:29.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:28 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:30.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: from='osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T13:55:30.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T13:55:30.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:30.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:30.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: pgmap v27: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T13:55:30.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2676151833' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5f2add4f-ea51-4973-85a8-19e508a69ff6"}]: dispatch 2026-03-09T13:55:30.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325] boot 2026-03-09T13:55:30.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2676151833' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5f2add4f-ea51-4973-85a8-19e508a69ff6"}]': finished 2026-03-09T13:55:30.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: osdmap e17: 4 total, 3 up, 4 in 2026-03-09T13:55:30.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:30.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:30.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:29 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3449107768' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: from='osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: osdmap e16: 3 total, 2 up, 3 in 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: pgmap v27: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2676151833' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5f2add4f-ea51-4973-85a8-19e508a69ff6"}]: dispatch 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: osd.2 [v2:192.168.123.108:6818/2187772325,v1:192.168.123.108:6819/2187772325] boot 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2676151833' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5f2add4f-ea51-4973-85a8-19e508a69ff6"}]': finished 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: osdmap e17: 4 total, 3 up, 4 in 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:30.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:29 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3449107768' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:30 vm10 ceph-mon[51861]: purged_snaps scrub starts 2026-03-09T13:55:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:30 vm10 ceph-mon[51861]: purged_snaps scrub ok 2026-03-09T13:55:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:30 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:30 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:30 vm10 ceph-mon[51861]: osdmap e18: 4 total, 3 up, 4 in 2026-03-09T13:55:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:30 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:30 vm08 ceph-mon[47734]: purged_snaps scrub starts 2026-03-09T13:55:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:30 vm08 ceph-mon[47734]: purged_snaps scrub ok 2026-03-09T13:55:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:30 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:30 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:30 vm08 ceph-mon[47734]: osdmap e18: 4 total, 3 up, 4 in 2026-03-09T13:55:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:30 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:32.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:31 vm08 ceph-mon[47734]: pgmap v30: 0 pgs: ; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T13:55:32.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:31 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]: dispatch 2026-03-09T13:55:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:31 vm10 ceph-mon[51861]: pgmap v30: 0 pgs: ; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T13:55:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:31 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]: dispatch 2026-03-09T13:55:33.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:32 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]': finished 2026-03-09T13:55:33.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:32 vm08 ceph-mon[47734]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T13:55:33.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:32 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:33.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:32 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T13:55:33.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:32 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T13:55:33.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:32 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:33.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:32 vm08 ceph-mon[47734]: Deploying daemon osd.3 on vm08 2026-03-09T13:55:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:32 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]': finished 2026-03-09T13:55:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:32 vm10 ceph-mon[51861]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T13:55:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:32 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:32 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T13:55:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:32 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T13:55:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:32 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:32 vm10 ceph-mon[51861]: Deploying daemon osd.3 on vm08 2026-03-09T13:55:34.253 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:33 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T13:55:34.253 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:33 vm08 ceph-mon[47734]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T13:55:34.253 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:33 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:34.253 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:33 vm08 ceph-mon[47734]: pgmap v33: 1 pgs: 1 creating+peering; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T13:55:34.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:33 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T13:55:34.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:33 vm10 ceph-mon[51861]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T13:55:34.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:33 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:34.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:33 vm10 ceph-mon[51861]: pgmap v33: 1 pgs: 1 creating+peering; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T13:55:35.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:34 vm08 ceph-mon[47734]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T13:55:35.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:35.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:35.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:35.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:35.037 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:34 vm10 ceph-mon[51861]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T13:55:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:35.635 INFO:journalctl@ceph.osd.0.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67793]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vde 2026-03-09T13:55:35.635 INFO:journalctl@ceph.osd.0.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67793]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T13:55:35.635 INFO:journalctl@ceph.osd.0.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67793]: pam_unix(sudo:session): session closed for user root 2026-03-09T13:55:35.687 INFO:teuthology.orchestra.run.vm08.stdout:Created osd(s) 3 on host 'vm08' 2026-03-09T13:55:35.744 DEBUG:teuthology.orchestra.run.vm08:osd.3> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@osd.3.service 2026-03-09T13:55:35.752 INFO:tasks.cephadm:Deploying osd.4 on vm10 with /dev/vde... 2026-03-09T13:55:35.752 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- lvm zap /dev/vde 2026-03-09T13:55:35.906 INFO:journalctl@ceph.osd.1.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67895]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vdd 2026-03-09T13:55:35.906 INFO:journalctl@ceph.osd.1.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67895]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T13:55:35.906 INFO:journalctl@ceph.osd.1.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67895]: pam_unix(sudo:session): session closed for user root 2026-03-09T13:55:36.044 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:35 vm10 ceph-mon[51861]: pgmap v35: 1 pgs: 1 creating+peering; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T13:55:36.044 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:35 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:36.172 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:35 vm08 ceph-mon[47734]: pgmap v35: 1 pgs: 1 creating+peering; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T13:55:36.173 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:35 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:36.173 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67945]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vdc 2026-03-09T13:55:36.173 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67945]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T13:55:36.173 INFO:journalctl@ceph.osd.2.vm08.stdout:Mar 09 13:55:35 vm08 sudo[67945]: pam_unix(sudo:session): session closed for user root 2026-03-09T13:55:36.378 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:55:36.391 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch daemon add osd vm10:/dev/vde 2026-03-09T13:55:36.465 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 sudo[68195]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vda 2026-03-09T13:55:36.465 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 sudo[68195]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T13:55:36.465 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 sudo[68195]: pam_unix(sudo:session): session closed for user root 2026-03-09T13:55:36.466 INFO:journalctl@ceph.osd.3.vm08.stdout:Mar 09 13:55:36 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-3[67384]: 2026-03-09T13:55:36.280+0000 7f91644f63c0 -1 osd.3 0 log_to_monitors true 2026-03-09T13:55:36.597 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 sudo[54204]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vda 2026-03-09T13:55:36.598 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 sudo[54204]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T13:55:36.598 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 sudo[54204]: pam_unix(sudo:session): session closed for user root 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:37.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm08"}]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:37.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:37.936 INFO:journalctl@ceph.osd.3.vm08.stdout:Mar 09 13:55:37 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-3[67384]: 2026-03-09T13:55:37.600+0000 7f915aef9700 -1 osd.3 0 waiting for initial osdmap 2026-03-09T13:55:37.936 INFO:journalctl@ceph.osd.3.vm08.stdout:Mar 09 13:55:37 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-3[67384]: 2026-03-09T13:55:37.609+0000 7f915508f700 -1 osd.3 23 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='client.24157 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: pgmap v36: 1 pgs: 1 creating+peering; 0 B data, 16 MiB used, 60 GiB / 60 GiB avail 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: osdmap e22: 4 total, 3 up, 4 in 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: mgrmap e19: vm08.siqces(active, since 54s), standbys: vm10.euefis 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/3293570328' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "78f72c80-39e6-4619-837c-3680a4d12223"}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/3293570328' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "78f72c80-39e6-4619-837c-3680a4d12223"}]': finished 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: osdmap e23: 5 total, 3 up, 5 in 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:38.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='client.24157 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: pgmap v36: 1 pgs: 1 creating+peering; 0 B data, 16 MiB used, 60 GiB / 60 GiB avail 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: osdmap e22: 4 total, 3 up, 4 in 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm08", "root=default"]}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: mgrmap e19: vm08.siqces(active, since 54s), standbys: vm10.euefis 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/3293570328' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "78f72c80-39e6-4619-837c-3680a4d12223"}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm08", "root=default"]}]': finished 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/3293570328' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "78f72c80-39e6-4619-837c-3680a4d12223"}]': finished 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: osdmap e23: 5 total, 3 up, 5 in 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:38.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:38 vm08 ceph-mon[47734]: Detected new or changed devices on vm08 2026-03-09T13:55:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:38 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/2290857575' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:39.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:38 vm08 ceph-mon[47734]: osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184] boot 2026-03-09T13:55:39.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:38 vm08 ceph-mon[47734]: osdmap e24: 5 total, 4 up, 5 in 2026-03-09T13:55:39.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:38 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:39.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:38 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:38 vm10 ceph-mon[51861]: Detected new or changed devices on vm08 2026-03-09T13:55:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:38 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/2290857575' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:38 vm10 ceph-mon[51861]: osd.3 [v2:192.168.123.108:6826/1203678184,v1:192.168.123.108:6827/1203678184] boot 2026-03-09T13:55:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:38 vm10 ceph-mon[51861]: osdmap e24: 5 total, 4 up, 5 in 2026-03-09T13:55:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:38 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T13:55:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:38 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:40.301 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:39 vm10 ceph-mon[51861]: purged_snaps scrub starts 2026-03-09T13:55:40.301 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:39 vm10 ceph-mon[51861]: purged_snaps scrub ok 2026-03-09T13:55:40.301 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:39 vm10 ceph-mon[51861]: pgmap v40: 1 pgs: 1 active+clean; 449 KiB data, 22 MiB used, 80 GiB / 80 GiB avail 2026-03-09T13:55:40.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:39 vm08 ceph-mon[47734]: purged_snaps scrub starts 2026-03-09T13:55:40.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:39 vm08 ceph-mon[47734]: purged_snaps scrub ok 2026-03-09T13:55:40.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:39 vm08 ceph-mon[47734]: pgmap v40: 1 pgs: 1 active+clean; 449 KiB data, 22 MiB used, 80 GiB / 80 GiB avail 2026-03-09T13:55:41.208 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:40 vm10 ceph-mon[51861]: osdmap e25: 5 total, 4 up, 5 in 2026-03-09T13:55:41.208 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:40 vm08 ceph-mon[47734]: osdmap e25: 5 total, 4 up, 5 in 2026-03-09T13:55:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:42.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:41 vm10 ceph-mon[51861]: pgmap v42: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T13:55:42.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T13:55:42.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:42.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:41 vm10 ceph-mon[51861]: Deploying daemon osd.4 on vm10 2026-03-09T13:55:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:41 vm08 ceph-mon[47734]: pgmap v42: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T13:55:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T13:55:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:41 vm08 ceph-mon[47734]: Deploying daemon osd.4 on vm10 2026-03-09T13:55:43.349 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 4 on host 'vm10' 2026-03-09T13:55:43.404 DEBUG:teuthology.orchestra.run.vm10:osd.4> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@osd.4.service 2026-03-09T13:55:43.406 INFO:tasks.cephadm:Deploying osd.5 on vm10 with /dev/vdd... 2026-03-09T13:55:43.406 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- lvm zap /dev/vdd 2026-03-09T13:55:43.542 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:43.542 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:43.542 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:43.542 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:43.542 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:55:43.542 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:55:43.542 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:43.542 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:55:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:55:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:44.092 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:55:44.107 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch daemon add osd vm10:/dev/vdd 2026-03-09T13:55:44.660 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:44 vm10 ceph-mon[51861]: pgmap v43: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T13:55:44.660 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:44 vm10 ceph-mon[51861]: from='osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T13:55:44.660 INFO:journalctl@ceph.osd.4.vm10.stdout:Mar 09 13:55:44 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-4[55818]: 2026-03-09T13:55:44.412+0000 7f6112ffc3c0 -1 osd.4 0 log_to_monitors true 2026-03-09T13:55:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:44 vm08 ceph-mon[47734]: pgmap v43: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T13:55:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:44 vm08 ceph-mon[47734]: from='osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:45.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:45.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:45.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/203076719' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85ed5c36-de18-4ce7-a2c8-bd4f4231e237"}]: dispatch 2026-03-09T13:55:45.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85ed5c36-de18-4ce7-a2c8-bd4f4231e237"}]: dispatch 2026-03-09T13:55:45.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T13:55:45.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "85ed5c36-de18-4ce7-a2c8-bd4f4231e237"}]': finished 2026-03-09T13:55:45.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: osdmap e27: 6 total, 4 up, 6 in 2026-03-09T13:55:45.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:45.553 INFO:journalctl@ceph.osd.4.vm10.stdout:Mar 09 13:55:45 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-4[55818]: 2026-03-09T13:55:45.460+0000 7f610b202700 -1 osd.4 0 waiting for initial osdmap 2026-03-09T13:55:45.553 INFO:journalctl@ceph.osd.4.vm10.stdout:Mar 09 13:55:45 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-4[55818]: 2026-03-09T13:55:45.470+0000 7f6103b95700 -1 osd.4 27 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/203076719' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85ed5c36-de18-4ce7-a2c8-bd4f4231e237"}]: dispatch 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85ed5c36-de18-4ce7-a2c8-bd4f4231e237"}]: dispatch 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "85ed5c36-de18-4ce7-a2c8-bd4f4231e237"}]': finished 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: osdmap e27: 6 total, 4 up, 6 in 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:46 vm08 ceph-mon[47734]: from='client.24175 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:46 vm08 ceph-mon[47734]: pgmap v45: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T13:55:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:46 vm08 ceph-mon[47734]: Detected new or changed devices on vm10 2026-03-09T13:55:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:46 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/131704744' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:46 vm08 ceph-mon[47734]: osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545] boot 2026-03-09T13:55:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:46 vm08 ceph-mon[47734]: osdmap e28: 6 total, 5 up, 6 in 2026-03-09T13:55:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:46 vm10 ceph-mon[51861]: from='client.24175 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:46 vm10 ceph-mon[51861]: pgmap v45: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T13:55:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:46 vm10 ceph-mon[51861]: Detected new or changed devices on vm10 2026-03-09T13:55:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:46 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/131704744' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:46 vm10 ceph-mon[51861]: osd.4 [v2:192.168.123.110:6800/3515807545,v1:192.168.123.110:6801/3515807545] boot 2026-03-09T13:55:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:46 vm10 ceph-mon[51861]: osdmap e28: 6 total, 5 up, 6 in 2026-03-09T13:55:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T13:55:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:47.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:47 vm10 ceph-mon[51861]: purged_snaps scrub starts 2026-03-09T13:55:47.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:47 vm10 ceph-mon[51861]: purged_snaps scrub ok 2026-03-09T13:55:47.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:47 vm10 ceph-mon[51861]: osdmap e29: 6 total, 5 up, 6 in 2026-03-09T13:55:47.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:47 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:48.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:47 vm08 ceph-mon[47734]: purged_snaps scrub starts 2026-03-09T13:55:48.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:47 vm08 ceph-mon[47734]: purged_snaps scrub ok 2026-03-09T13:55:48.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:47 vm08 ceph-mon[47734]: osdmap e29: 6 total, 5 up, 6 in 2026-03-09T13:55:48.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:47 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:48.773 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:48 vm10 ceph-mon[51861]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T13:55:48.773 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:48 vm10 ceph-mon[51861]: osdmap e30: 6 total, 5 up, 6 in 2026-03-09T13:55:48.773 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:48 vm08 ceph-mon[47734]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T13:55:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:48 vm08 ceph-mon[47734]: osdmap e30: 6 total, 5 up, 6 in 2026-03-09T13:55:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:49.620 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T13:55:49.620 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T13:55:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:50.686 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:50 vm10 ceph-mon[51861]: Deploying daemon osd.5 on vm10 2026-03-09T13:55:50.686 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:50 vm10 ceph-mon[51861]: pgmap v51: 1 pgs: 1 peering; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T13:55:50.687 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:50.687 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:50.687 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:50.687 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:50 vm08 ceph-mon[47734]: Deploying daemon osd.5 on vm10 2026-03-09T13:55:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:50 vm08 ceph-mon[47734]: pgmap v51: 1 pgs: 1 peering; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T13:55:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:51.222 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 5 on host 'vm10' 2026-03-09T13:55:51.284 DEBUG:teuthology.orchestra.run.vm10:osd.5> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@osd.5.service 2026-03-09T13:55:51.286 INFO:tasks.cephadm:Deploying osd.6 on vm10 with /dev/vdc... 2026-03-09T13:55:51.286 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- lvm zap /dev/vdc 2026-03-09T13:55:51.909 INFO:journalctl@ceph.osd.5.vm10.stdout:Mar 09 13:55:51 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-5[58664]: 2026-03-09T13:55:51.742+0000 7fa0e03b63c0 -1 osd.5 0 log_to_monitors true 2026-03-09T13:55:52.524 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:52 vm10 ceph-mon[51861]: pgmap v52: 1 pgs: 1 peering; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T13:55:52.524 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:52.524 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:52.524 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:52 vm10 ceph-mon[51861]: from='osd.5 [v2:192.168.123.110:6808/1405421093,v1:192.168.123.110:6809/1405421093]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T13:55:52.524 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:52 vm10 ceph-mon[51861]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T13:55:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:52 vm08 ceph-mon[47734]: pgmap v52: 1 pgs: 1 peering; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T13:55:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:52 vm08 ceph-mon[47734]: from='osd.5 [v2:192.168.123.110:6808/1405421093,v1:192.168.123.110:6809/1405421093]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T13:55:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:52 vm08 ceph-mon[47734]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T13:55:52.873 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:55:52.888 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch daemon add osd vm10:/dev/vdc 2026-03-09T13:55:53.658 INFO:journalctl@ceph.osd.5.vm10.stdout:Mar 09 13:55:53 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-5[58664]: 2026-03-09T13:55:53.587+0000 7fa0d6db9700 -1 osd.5 0 waiting for initial osdmap 2026-03-09T13:55:53.658 INFO:journalctl@ceph.osd.5.vm10.stdout:Mar 09 13:55:53 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-5[58664]: 2026-03-09T13:55:53.603+0000 7fa0d0f4f700 -1 osd.5 32 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: Detected new or changed devices on vm10 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='osd.5 [v2:192.168.123.110:6808/1405421093,v1:192.168.123.110:6809/1405421093]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:53.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: Detected new or changed devices on vm10 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='osd.5 [v2:192.168.123.110:6808/1405421093,v1:192.168.123.110:6809/1405421093]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:55:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:53 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: pgmap v54: 1 pgs: 1 active+recovering; 449 KiB data, 29 MiB used, 100 GiB / 100 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: from='client.24193 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: osdmap e32: 6 total, 5 up, 6 in 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/4131883241' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6f33888e-59b3-4b64-acec-390be180037c"}]: dispatch 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: osd.5 [v2:192.168.123.110:6808/1405421093,v1:192.168.123.110:6809/1405421093] boot 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/4131883241' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6f33888e-59b3-4b64-acec-390be180037c"}]': finished 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: osdmap e33: 7 total, 6 up, 7 in 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:54.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:54 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:55:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: pgmap v54: 1 pgs: 1 active+recovering; 449 KiB data, 29 MiB used, 100 GiB / 100 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-09T13:55:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: from='client.24193 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: osdmap e32: 6 total, 5 up, 6 in 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/4131883241' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6f33888e-59b3-4b64-acec-390be180037c"}]: dispatch 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: osd.5 [v2:192.168.123.110:6808/1405421093,v1:192.168.123.110:6809/1405421093] boot 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/4131883241' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6f33888e-59b3-4b64-acec-390be180037c"}]': finished 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: osdmap e33: 7 total, 6 up, 7 in 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T13:55:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:54 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:55:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:55 vm10 ceph-mon[51861]: purged_snaps scrub starts 2026-03-09T13:55:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:55 vm10 ceph-mon[51861]: purged_snaps scrub ok 2026-03-09T13:55:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:55 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/1328996370' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:55.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:55 vm10 ceph-mon[51861]: osdmap e34: 7 total, 6 up, 7 in 2026-03-09T13:55:55.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:55 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:55:56.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:55 vm08 ceph-mon[47734]: purged_snaps scrub starts 2026-03-09T13:55:56.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:55 vm08 ceph-mon[47734]: purged_snaps scrub ok 2026-03-09T13:55:56.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:55 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/1328996370' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:55:56.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:55 vm08 ceph-mon[47734]: osdmap e34: 7 total, 6 up, 7 in 2026-03-09T13:55:56.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:55 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:55:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:56 vm08 ceph-mon[47734]: pgmap v58: 1 pgs: 1 active+recovering; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail; 97 KiB/s, 0 objects/s recovering 2026-03-09T13:55:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:56 vm08 ceph-mon[47734]: osdmap e35: 7 total, 6 up, 7 in 2026-03-09T13:55:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:56 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:55:56.856 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:56 vm10 ceph-mon[51861]: pgmap v58: 1 pgs: 1 active+recovering; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail; 97 KiB/s, 0 objects/s recovering 2026-03-09T13:55:56.856 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:56 vm10 ceph-mon[51861]: osdmap e35: 7 total, 6 up, 7 in 2026-03-09T13:55:56.856 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:56 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:55:57.867 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:57 vm10 ceph-mon[51861]: pgmap v60: 1 pgs: 1 active+recovering; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T13:55:58.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:57 vm08 ceph-mon[47734]: pgmap v60: 1 pgs: 1 active+recovering; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T13:55:58.678 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:58 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-09T13:55:58.678 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:58 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:58.678 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:55:58 vm10 ceph-mon[51861]: Deploying daemon osd.6 on vm10 2026-03-09T13:55:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:58 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-09T13:55:59.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:58 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:55:59.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:55:58 vm08 ceph-mon[47734]: Deploying daemon osd.6 on vm10 2026-03-09T13:56:00.588 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 6 on host 'vm10' 2026-03-09T13:56:00.626 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:00 vm10 ceph-mon[51861]: pgmap v61: 1 pgs: 1 active+recovering; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T13:56:00.626 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:00.626 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:00.626 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:00.626 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:00 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:00.661 DEBUG:teuthology.orchestra.run.vm10:osd.6> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@osd.6.service 2026-03-09T13:56:00.662 INFO:tasks.cephadm:Deploying osd.7 on vm10 with /dev/vdb... 2026-03-09T13:56:00.662 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- lvm zap /dev/vdb 2026-03-09T13:56:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:00 vm08 ceph-mon[47734]: pgmap v61: 1 pgs: 1 active+recovering; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T13:56:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:00 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:00.901 INFO:journalctl@ceph.osd.6.vm10.stdout:Mar 09 13:56:00 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-6[61460]: 2026-03-09T13:56:00.785+0000 7fdffe4ca3c0 -1 osd.6 0 log_to_monitors true 2026-03-09T13:56:01.507 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T13:56:01.529 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch daemon add osd vm10:/dev/vdb 2026-03-09T13:56:01.821 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:01.821 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:01 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:01.821 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:01 vm10 ceph-mon[51861]: from='osd.6 [v2:192.168.123.110:6816/1648093007,v1:192.168.123.110:6817/1648093007]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T13:56:01.821 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:01 vm10 ceph-mon[51861]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T13:56:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:01 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:01 vm08 ceph-mon[47734]: from='osd.6 [v2:192.168.123.110:6816/1648093007,v1:192.168.123.110:6817/1648093007]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T13:56:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:01 vm08 ceph-mon[47734]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T13:56:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: pgmap v62: 1 pgs: 1 active+recovering; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-09T13:56:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: osdmap e36: 7 total, 6 up, 7 in 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='osd.6 [v2:192.168.123.110:6816/1648093007,v1:192.168.123.110:6817/1648093007]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='client.24211 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: Detected new or changed devices on vm10 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:02.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:02 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:02.909 INFO:journalctl@ceph.osd.6.vm10.stdout:Mar 09 13:56:02 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-6[61460]: 2026-03-09T13:56:02.653+0000 7fdff66d0700 -1 osd.6 0 waiting for initial osdmap 2026-03-09T13:56:02.909 INFO:journalctl@ceph.osd.6.vm10.stdout:Mar 09 13:56:02 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-6[61460]: 2026-03-09T13:56:02.665+0000 7fdfef063700 -1 osd.6 37 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T13:56:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: pgmap v62: 1 pgs: 1 active+recovering; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: osdmap e36: 7 total, 6 up, 7 in 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='osd.6 [v2:192.168.123.110:6816/1648093007,v1:192.168.123.110:6817/1648093007]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='osd.6 ' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='client.24211 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: Detected new or changed devices on vm10 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:02 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: osdmap e37: 7 total, 6 up, 7 in 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: pgmap v65: 1 pgs: 1 active+recovering; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/2839955714' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6010e654-8935-42d2-9f1d-5fb78d19a935"}]: dispatch 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6010e654-8935-42d2-9f1d-5fb78d19a935"}]: dispatch 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: osd.6 [v2:192.168.123.110:6816/1648093007,v1:192.168.123.110:6817/1648093007] boot 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6010e654-8935-42d2-9f1d-5fb78d19a935"}]': finished 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: osdmap e38: 8 total, 7 up, 8 in 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:03.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:03 vm08 ceph-mon[47734]: from='client.? 192.168.123.110:0/3822883807' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: osdmap e37: 7 total, 6 up, 7 in 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: pgmap v65: 1 pgs: 1 active+recovering; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/2839955714' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6010e654-8935-42d2-9f1d-5fb78d19a935"}]: dispatch 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6010e654-8935-42d2-9f1d-5fb78d19a935"}]: dispatch 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: osd.6 [v2:192.168.123.110:6816/1648093007,v1:192.168.123.110:6817/1648093007] boot 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6010e654-8935-42d2-9f1d-5fb78d19a935"}]': finished 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: osdmap e38: 8 total, 7 up, 8 in 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T13:56:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:03.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:03 vm10 ceph-mon[51861]: from='client.? 192.168.123.110:0/3822883807' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T13:56:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:04 vm08 ceph-mon[47734]: purged_snaps scrub starts 2026-03-09T13:56:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:04 vm08 ceph-mon[47734]: purged_snaps scrub ok 2026-03-09T13:56:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:04 vm08 ceph-mon[47734]: osdmap e39: 8 total, 7 up, 8 in 2026-03-09T13:56:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:04 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:04 vm10 ceph-mon[51861]: purged_snaps scrub starts 2026-03-09T13:56:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:04 vm10 ceph-mon[51861]: purged_snaps scrub ok 2026-03-09T13:56:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:04 vm10 ceph-mon[51861]: osdmap e39: 8 total, 7 up, 8 in 2026-03-09T13:56:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:04 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:06.085 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:06 vm10 ceph-mon[51861]: pgmap v68: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail 2026-03-09T13:56:06.085 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:06 vm10 ceph-mon[51861]: osdmap e40: 8 total, 7 up, 8 in 2026-03-09T13:56:06.085 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:06 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:06 vm08 ceph-mon[47734]: pgmap v68: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail 2026-03-09T13:56:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:06 vm08 ceph-mon[47734]: osdmap e40: 8 total, 7 up, 8 in 2026-03-09T13:56:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:06 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:07 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-09T13:56:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:07 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:07 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-09T13:56:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:07 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:08.351 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:08 vm10 ceph-mon[51861]: Deploying daemon osd.7 on vm10 2026-03-09T13:56:08.351 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:08 vm10 ceph-mon[51861]: pgmap v70: 1 pgs: 1 active+recovering; 449 KiB data, 40 MiB used, 140 GiB / 140 GiB avail 2026-03-09T13:56:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:08 vm08 ceph-mon[47734]: Deploying daemon osd.7 on vm10 2026-03-09T13:56:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:08 vm08 ceph-mon[47734]: pgmap v70: 1 pgs: 1 active+recovering; 449 KiB data, 40 MiB used, 140 GiB / 140 GiB avail 2026-03-09T13:56:09.195 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:09 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:09.195 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:09 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:09.195 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:09 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:09.195 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:09 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:09 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:09 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:09 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:09 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:09.592 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 7 on host 'vm10' 2026-03-09T13:56:09.665 DEBUG:teuthology.orchestra.run.vm10:osd.7> sudo journalctl -f -n 0 -u ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64@osd.7.service 2026-03-09T13:56:09.667 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-09T13:56:09.667 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd stat -f json 2026-03-09T13:56:10.152 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:10.194 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":40,"num_osds":8,"num_up_osds":7,"osd_up_since":1773064563,"num_in_osds":8,"osd_in_since":1773064563,"num_remapped_pgs":0} 2026-03-09T13:56:10.408 INFO:journalctl@ceph.osd.7.vm10.stdout:Mar 09 13:56:10 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-7[64271]: 2026-03-09T13:56:10.122+0000 7fca0115a3c0 -1 osd.7 0 log_to_monitors true 2026-03-09T13:56:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:10 vm10 ceph-mon[51861]: pgmap v71: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail 2026-03-09T13:56:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:10 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:10 vm10 ceph-mon[51861]: from='osd.7 [v2:192.168.123.110:6824/303050948,v1:192.168.123.110:6825/303050948]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T13:56:10.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:10 vm10 ceph-mon[51861]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T13:56:10.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:10 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/1755000179' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T13:56:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:10 vm08 ceph-mon[47734]: pgmap v71: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail 2026-03-09T13:56:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:10 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:10 vm08 ceph-mon[47734]: from='osd.7 [v2:192.168.123.110:6824/303050948,v1:192.168.123.110:6825/303050948]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T13:56:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:10 vm08 ceph-mon[47734]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T13:56:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:10 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1755000179' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T13:56:11.195 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd stat -f json 2026-03-09T13:56:11.640 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:11.691 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":41,"num_osds":8,"num_up_osds":7,"osd_up_since":1773064563,"num_in_osds":8,"osd_in_since":1773064563,"num_remapped_pgs":0} 2026-03-09T13:56:12.158 INFO:journalctl@ceph.osd.7.vm10.stdout:Mar 09 13:56:11 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-7[64271]: 2026-03-09T13:56:11.879+0000 7fc9f7b5d700 -1 osd.7 0 waiting for initial osdmap 2026-03-09T13:56:12.158 INFO:journalctl@ceph.osd.7.vm10.stdout:Mar 09 13:56:11 vm10 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-osd-7[64271]: 2026-03-09T13:56:11.891+0000 7fc9f1cf3700 -1 osd.7 42 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: osdmap e41: 8 total, 7 up, 8 in 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='osd.7 [v2:192.168.123.110:6824/303050948,v1:192.168.123.110:6825/303050948]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: pgmap v73: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: Detected new or changed devices on vm10 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:56:12.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:12.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:12.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:12.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:12.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:12.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:11 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2422059873' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: osdmap e41: 8 total, 7 up, 8 in 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='osd.7 [v2:192.168.123.110:6824/303050948,v1:192.168.123.110:6825/303050948]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: pgmap v73: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: Detected new or changed devices on vm10 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:12.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:11 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2422059873' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T13:56:12.692 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd stat -f json 2026-03-09T13:56:12.970 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:12 vm08 ceph-mon[47734]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T13:56:12.970 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:12 vm08 ceph-mon[47734]: osdmap e42: 8 total, 7 up, 8 in 2026-03-09T13:56:12.971 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:12 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:12.971 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:12 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:12.971 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:12 vm08 ceph-mon[47734]: osd.7 [v2:192.168.123.110:6824/303050948,v1:192.168.123.110:6825/303050948] boot 2026-03-09T13:56:12.971 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:12 vm08 ceph-mon[47734]: osdmap e43: 8 total, 8 up, 8 in 2026-03-09T13:56:12.971 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:12 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:12 vm10 ceph-mon[51861]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T13:56:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:12 vm10 ceph-mon[51861]: osdmap e42: 8 total, 7 up, 8 in 2026-03-09T13:56:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:12 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:12 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:12 vm10 ceph-mon[51861]: osd.7 [v2:192.168.123.110:6824/303050948,v1:192.168.123.110:6825/303050948] boot 2026-03-09T13:56:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:12 vm10 ceph-mon[51861]: osdmap e43: 8 total, 8 up, 8 in 2026-03-09T13:56:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:12 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T13:56:13.186 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:13.258 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":43,"num_osds":8,"num_up_osds":8,"osd_up_since":1773064572,"num_in_osds":8,"osd_in_since":1773064563,"num_remapped_pgs":1} 2026-03-09T13:56:13.258 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd dump --format=json 2026-03-09T13:56:13.408 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:13.724 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:13.724 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":43,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","created":"2026-03-09T13:53:32.661723+0000","modified":"2026-03-09T13:56:12.868921+0000","last_up_change":"2026-03-09T13:56:12.868921+0000","last_in_change":"2026-03-09T13:56:03.071679+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T13:55:31.113056+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"8426b89c-8bdd-425e-8381-20b91a25bd3d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":42,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6803","nonce":2288390321}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6805","nonce":2288390321}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6809","nonce":2288390321}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6807","nonce":2288390321}]},"public_addr":"192.168.123.108:6803/2288390321","cluster_addr":"192.168.123.108:6805/2288390321","heartbeat_back_addr":"192.168.123.108:6809/2288390321","heartbeat_front_addr":"192.168.123.108:6807/2288390321","state":["exists","up"]},{"osd":1,"uuid":"7759a684-08a6-4323-b3cb-092ae4decfe1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":29,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6811","nonce":3586975833}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6813","nonce":3586975833}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6817","nonce":3586975833}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6815","nonce":3586975833}]},"public_addr":"192.168.123.108:6811/3586975833","cluster_addr":"192.168.123.108:6813/3586975833","heartbeat_back_addr":"192.168.123.108:6817/3586975833","heartbeat_front_addr":"192.168.123.108:6815/3586975833","state":["exists","up"]},{"osd":2,"uuid":"8a88652c-970b-4ebe-8175-ea8b9c059735","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6819","nonce":2187772325}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6821","nonce":2187772325}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6825","nonce":2187772325}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6823","nonce":2187772325}]},"public_addr":"192.168.123.108:6819/2187772325","cluster_addr":"192.168.123.108:6821/2187772325","heartbeat_back_addr":"192.168.123.108:6825/2187772325","heartbeat_front_addr":"192.168.123.108:6823/2187772325","state":["exists","up"]},{"osd":3,"uuid":"5f2add4f-ea51-4973-85a8-19e508a69ff6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":24,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6827","nonce":1203678184}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6829","nonce":1203678184}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6832","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6833","nonce":1203678184}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6831","nonce":1203678184}]},"public_addr":"192.168.123.108:6827/1203678184","cluster_addr":"192.168.123.108:6829/1203678184","heartbeat_back_addr":"192.168.123.108:6833/1203678184","heartbeat_front_addr":"192.168.123.108:6831/1203678184","state":["exists","up"]},{"osd":4,"uuid":"78f72c80-39e6-4619-837c-3680a4d12223","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":28,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6801","nonce":3515807545}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6803","nonce":3515807545}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6807","nonce":3515807545}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6805","nonce":3515807545}]},"public_addr":"192.168.123.110:6801/3515807545","cluster_addr":"192.168.123.110:6803/3515807545","heartbeat_back_addr":"192.168.123.110:6807/3515807545","heartbeat_front_addr":"192.168.123.110:6805/3515807545","state":["exists","up"]},{"osd":5,"uuid":"85ed5c36-de18-4ce7-a2c8-bd4f4231e237","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":33,"up_thru":34,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6809","nonce":1405421093}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6811","nonce":1405421093}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6815","nonce":1405421093}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6813","nonce":1405421093}]},"public_addr":"192.168.123.110:6809/1405421093","cluster_addr":"192.168.123.110:6811/1405421093","heartbeat_back_addr":"192.168.123.110:6815/1405421093","heartbeat_front_addr":"192.168.123.110:6813/1405421093","state":["exists","up"]},{"osd":6,"uuid":"6f33888e-59b3-4b64-acec-390be180037c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":38,"up_thru":39,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6816","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6817","nonce":1648093007}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6818","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6819","nonce":1648093007}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6822","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6823","nonce":1648093007}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6820","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6821","nonce":1648093007}]},"public_addr":"192.168.123.110:6817/1648093007","cluster_addr":"192.168.123.110:6819/1648093007","heartbeat_back_addr":"192.168.123.110:6823/1648093007","heartbeat_front_addr":"192.168.123.110:6821/1648093007","state":["exists","up"]},{"osd":7,"uuid":"6010e654-8935-42d2-9f1d-5fb78d19a935","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":43,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6824","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6825","nonce":303050948}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6826","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6827","nonce":303050948}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6830","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6831","nonce":303050948}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6828","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6829","nonce":303050948}]},"public_addr":"192.168.123.110:6825/303050948","cluster_addr":"192.168.123.110:6827/303050948","heartbeat_back_addr":"192.168.123.110:6831/303050948","heartbeat_front_addr":"192.168.123.110:6829/303050948","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:10.341306+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:20.310352+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:28.227983+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:37.316803+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:45.441537+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:52.730521+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:56:01.814432+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[{"pgid":"1.0","osds":[0,6,1]}],"primary_temp":[],"blocklist":{"192.168.123.108:6800/1786960600":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6801/1786960600":"2026-03-10T13:54:42.843895+0000","192.168.123.108:0/2425398216":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6800/2399132916":"2026-03-10T13:54:01.629885+0000","192.168.123.108:6801/2399132916":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/246681717":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/2203639542":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/3049100618":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/2992765302":"2026-03-10T13:54:42.843895+0000","192.168.123.108:0/489616954":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/3582363488":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/46790109":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6800/677438560":"2026-03-10T13:53:49.429024+0000","192.168.123.108:6801/677438560":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/2681069985":"2026-03-10T13:53:49.429024+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T13:56:13.783 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-09T13:55:31.113056+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '21', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}}] 2026-03-09T13:56:13.783 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd pool get .mgr pg_num 2026-03-09T13:56:13.946 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:13.977 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:13 vm08 ceph-mon[47734]: purged_snaps scrub starts 2026-03-09T13:56:13.977 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:13 vm08 ceph-mon[47734]: purged_snaps scrub ok 2026-03-09T13:56:13.977 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:13 vm08 ceph-mon[47734]: pgmap v76: 1 pgs: 1 active+recovering; 449 KiB data, 46 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:13.977 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:13 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/260648656' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T13:56:13.977 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:13 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/4089880181' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T13:56:13.977 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:13 vm08 ceph-mon[47734]: osdmap e44: 8 total, 8 up, 8 in 2026-03-09T13:56:14.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:13 vm10 ceph-mon[51861]: purged_snaps scrub starts 2026-03-09T13:56:14.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:13 vm10 ceph-mon[51861]: purged_snaps scrub ok 2026-03-09T13:56:14.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:13 vm10 ceph-mon[51861]: pgmap v76: 1 pgs: 1 active+recovering; 449 KiB data, 46 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:14.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:13 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/260648656' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T13:56:14.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:13 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/4089880181' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T13:56:14.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:13 vm10 ceph-mon[51861]: osdmap e44: 8 total, 8 up, 8 in 2026-03-09T13:56:14.299 INFO:teuthology.orchestra.run.vm08.stdout:pg_num: 1 2026-03-09T13:56:14.341 INFO:tasks.cephadm:Setting up client nodes... 2026-03-09T13:56:14.341 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-09T13:56:14.803 INFO:teuthology.orchestra.run.vm08.stdout:[client.0] 2026-03-09T13:56:14.803 INFO:teuthology.orchestra.run.vm08.stdout: key = AQB+0a5ptNClLxAAarspdbAbj/iMN80YHs1RuA== 2026-03-09T13:56:14.879 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-03-09T13:56:14.879 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-09T13:56:14.879 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-09T13:56:14.918 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-09T13:56:14.918 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-09T13:56:14.918 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph mgr dump --format=json 2026-03-09T13:56:15.107 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:15.131 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:14 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2564455748' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T13:56:15.131 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:14 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3784884380' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T13:56:15.131 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:14 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3784884380' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T13:56:15.131 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:14 vm08 ceph-mon[47734]: osdmap e45: 8 total, 8 up, 8 in 2026-03-09T13:56:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:14 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2564455748' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T13:56:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:14 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3784884380' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T13:56:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:14 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3784884380' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T13:56:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:14 vm10 ceph-mon[51861]: osdmap e45: 8 total, 8 up, 8 in 2026-03-09T13:56:15.456 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:15.502 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":19,"active_gid":14232,"active_name":"vm08.siqces","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":473541395},{"type":"v1","addr":"192.168.123.108:6801","nonce":473541395}]},"active_addr":"192.168.123.108:6801/473541395","active_change":"2026-03-09T13:54:42.845205+0000","active_mgr_features":4540138303579357183,"available":true,"standbys":[{"gid":14250,"name":"vm10.euefis","mgr_features":4540138303579357183,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2400","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"7","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","upmap"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.23.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/ceph-grafana:8.3.5","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"docker.io/library/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"docker.io/arcts/keepalived","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.3.1","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.33.4","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"docker.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"noautoscale":{"name":"noautoscale","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"global autoscale flag","long_desc":"Option to turn on/off the autoscaler for all pools","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"drive_group_interval":{"name":"drive_group_interval","type":"float","level":"advanced","flags":0,"default_value":"300.0","min":"","max":"","enum_allowed":[],"desc":"interval in seconds between re-application of applied drive_groups","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2400","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"7","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","upmap"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.23.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/ceph-grafana:8.3.5","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"docker.io/library/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"docker.io/arcts/keepalived","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.3.1","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.33.4","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"docker.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"noautoscale":{"name":"noautoscale","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"global autoscale flag","long_desc":"Option to turn on/off the autoscaler for all pools","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"drive_group_interval":{"name":"drive_group_interval","type":"float","level":"advanced","flags":0,"default_value":"300.0","min":"","max":"","enum_allowed":[],"desc":"interval in seconds between re-application of applied drive_groups","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.108:8443/","prometheus":"http://192.168.123.108:9283/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"last_failure_osd_epoch":5,"active_clients":[{"addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":4261797738}]},{"addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":1023147594}]},{"addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":2068582384}]},{"addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":1383634589}]}]}} 2026-03-09T13:56:15.503 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-09T13:56:15.503 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-09T13:56:15.504 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd dump --format=json 2026-03-09T13:56:15.652 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:15.979 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:15.979 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":45,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","created":"2026-03-09T13:53:32.661723+0000","modified":"2026-03-09T13:56:14.873520+0000","last_up_change":"2026-03-09T13:56:12.868921+0000","last_in_change":"2026-03-09T13:56:03.071679+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T13:55:31.113056+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"8426b89c-8bdd-425e-8381-20b91a25bd3d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":43,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6803","nonce":2288390321}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6805","nonce":2288390321}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6809","nonce":2288390321}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6807","nonce":2288390321}]},"public_addr":"192.168.123.108:6803/2288390321","cluster_addr":"192.168.123.108:6805/2288390321","heartbeat_back_addr":"192.168.123.108:6809/2288390321","heartbeat_front_addr":"192.168.123.108:6807/2288390321","state":["exists","up"]},{"osd":1,"uuid":"7759a684-08a6-4323-b3cb-092ae4decfe1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":29,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6811","nonce":3586975833}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6813","nonce":3586975833}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6817","nonce":3586975833}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6815","nonce":3586975833}]},"public_addr":"192.168.123.108:6811/3586975833","cluster_addr":"192.168.123.108:6813/3586975833","heartbeat_back_addr":"192.168.123.108:6817/3586975833","heartbeat_front_addr":"192.168.123.108:6815/3586975833","state":["exists","up"]},{"osd":2,"uuid":"8a88652c-970b-4ebe-8175-ea8b9c059735","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6819","nonce":2187772325}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6821","nonce":2187772325}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6825","nonce":2187772325}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6823","nonce":2187772325}]},"public_addr":"192.168.123.108:6819/2187772325","cluster_addr":"192.168.123.108:6821/2187772325","heartbeat_back_addr":"192.168.123.108:6825/2187772325","heartbeat_front_addr":"192.168.123.108:6823/2187772325","state":["exists","up"]},{"osd":3,"uuid":"5f2add4f-ea51-4973-85a8-19e508a69ff6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":24,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6827","nonce":1203678184}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6829","nonce":1203678184}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6832","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6833","nonce":1203678184}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6831","nonce":1203678184}]},"public_addr":"192.168.123.108:6827/1203678184","cluster_addr":"192.168.123.108:6829/1203678184","heartbeat_back_addr":"192.168.123.108:6833/1203678184","heartbeat_front_addr":"192.168.123.108:6831/1203678184","state":["exists","up"]},{"osd":4,"uuid":"78f72c80-39e6-4619-837c-3680a4d12223","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":28,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6801","nonce":3515807545}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6803","nonce":3515807545}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6807","nonce":3515807545}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6805","nonce":3515807545}]},"public_addr":"192.168.123.110:6801/3515807545","cluster_addr":"192.168.123.110:6803/3515807545","heartbeat_back_addr":"192.168.123.110:6807/3515807545","heartbeat_front_addr":"192.168.123.110:6805/3515807545","state":["exists","up"]},{"osd":5,"uuid":"85ed5c36-de18-4ce7-a2c8-bd4f4231e237","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":33,"up_thru":34,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6809","nonce":1405421093}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6811","nonce":1405421093}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6815","nonce":1405421093}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6813","nonce":1405421093}]},"public_addr":"192.168.123.110:6809/1405421093","cluster_addr":"192.168.123.110:6811/1405421093","heartbeat_back_addr":"192.168.123.110:6815/1405421093","heartbeat_front_addr":"192.168.123.110:6813/1405421093","state":["exists","up"]},{"osd":6,"uuid":"6f33888e-59b3-4b64-acec-390be180037c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":38,"up_thru":39,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6816","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6817","nonce":1648093007}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6818","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6819","nonce":1648093007}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6822","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6823","nonce":1648093007}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6820","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6821","nonce":1648093007}]},"public_addr":"192.168.123.110:6817/1648093007","cluster_addr":"192.168.123.110:6819/1648093007","heartbeat_back_addr":"192.168.123.110:6823/1648093007","heartbeat_front_addr":"192.168.123.110:6821/1648093007","state":["exists","up"]},{"osd":7,"uuid":"6010e654-8935-42d2-9f1d-5fb78d19a935","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":43,"up_thru":44,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6824","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6825","nonce":303050948}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6826","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6827","nonce":303050948}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6830","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6831","nonce":303050948}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6828","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6829","nonce":303050948}]},"public_addr":"192.168.123.110:6825/303050948","cluster_addr":"192.168.123.110:6827/303050948","heartbeat_back_addr":"192.168.123.110:6831/303050948","heartbeat_front_addr":"192.168.123.110:6829/303050948","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:10.341306+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:20.310352+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:28.227983+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:37.316803+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:45.441537+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:52.730521+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:56:01.814432+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:56:11.151529+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.108:6800/1786960600":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6801/1786960600":"2026-03-10T13:54:42.843895+0000","192.168.123.108:0/2425398216":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6800/2399132916":"2026-03-10T13:54:01.629885+0000","192.168.123.108:6801/2399132916":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/246681717":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/2203639542":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/3049100618":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/2992765302":"2026-03-10T13:54:42.843895+0000","192.168.123.108:0/489616954":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/3582363488":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/46790109":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6800/677438560":"2026-03-10T13:53:49.429024+0000","192.168.123.108:6801/677438560":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/2681069985":"2026-03-10T13:53:49.429024+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T13:56:15.979 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:15 vm08 ceph-mon[47734]: pgmap v79: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:15.979 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:15 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/4132954190' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T13:56:16.052 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-09T13:56:16.052 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd dump --format=json 2026-03-09T13:56:16.193 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:16.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:15 vm10 ceph-mon[51861]: pgmap v79: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:16.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:15 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/4132954190' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T13:56:16.533 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:16.533 INFO:teuthology.orchestra.run.vm08.stdout:{"epoch":45,"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","created":"2026-03-09T13:53:32.661723+0000","modified":"2026-03-09T13:56:14.873520+0000","last_up_change":"2026-03-09T13:56:12.868921+0000","last_in_change":"2026-03-09T13:56:03.071679+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T13:55:31.113056+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"8426b89c-8bdd-425e-8381-20b91a25bd3d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":43,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6803","nonce":2288390321}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6805","nonce":2288390321}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6809","nonce":2288390321}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":2288390321},{"type":"v1","addr":"192.168.123.108:6807","nonce":2288390321}]},"public_addr":"192.168.123.108:6803/2288390321","cluster_addr":"192.168.123.108:6805/2288390321","heartbeat_back_addr":"192.168.123.108:6809/2288390321","heartbeat_front_addr":"192.168.123.108:6807/2288390321","state":["exists","up"]},{"osd":1,"uuid":"7759a684-08a6-4323-b3cb-092ae4decfe1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":29,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6811","nonce":3586975833}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6813","nonce":3586975833}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6817","nonce":3586975833}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":3586975833},{"type":"v1","addr":"192.168.123.108:6815","nonce":3586975833}]},"public_addr":"192.168.123.108:6811/3586975833","cluster_addr":"192.168.123.108:6813/3586975833","heartbeat_back_addr":"192.168.123.108:6817/3586975833","heartbeat_front_addr":"192.168.123.108:6815/3586975833","state":["exists","up"]},{"osd":2,"uuid":"8a88652c-970b-4ebe-8175-ea8b9c059735","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6819","nonce":2187772325}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6821","nonce":2187772325}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6825","nonce":2187772325}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":2187772325},{"type":"v1","addr":"192.168.123.108:6823","nonce":2187772325}]},"public_addr":"192.168.123.108:6819/2187772325","cluster_addr":"192.168.123.108:6821/2187772325","heartbeat_back_addr":"192.168.123.108:6825/2187772325","heartbeat_front_addr":"192.168.123.108:6823/2187772325","state":["exists","up"]},{"osd":3,"uuid":"5f2add4f-ea51-4973-85a8-19e508a69ff6","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":24,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6827","nonce":1203678184}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6829","nonce":1203678184}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6832","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6833","nonce":1203678184}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":1203678184},{"type":"v1","addr":"192.168.123.108:6831","nonce":1203678184}]},"public_addr":"192.168.123.108:6827/1203678184","cluster_addr":"192.168.123.108:6829/1203678184","heartbeat_back_addr":"192.168.123.108:6833/1203678184","heartbeat_front_addr":"192.168.123.108:6831/1203678184","state":["exists","up"]},{"osd":4,"uuid":"78f72c80-39e6-4619-837c-3680a4d12223","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":28,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6801","nonce":3515807545}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6803","nonce":3515807545}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6807","nonce":3515807545}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":3515807545},{"type":"v1","addr":"192.168.123.110:6805","nonce":3515807545}]},"public_addr":"192.168.123.110:6801/3515807545","cluster_addr":"192.168.123.110:6803/3515807545","heartbeat_back_addr":"192.168.123.110:6807/3515807545","heartbeat_front_addr":"192.168.123.110:6805/3515807545","state":["exists","up"]},{"osd":5,"uuid":"85ed5c36-de18-4ce7-a2c8-bd4f4231e237","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":33,"up_thru":34,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6809","nonce":1405421093}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6811","nonce":1405421093}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6815","nonce":1405421093}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":1405421093},{"type":"v1","addr":"192.168.123.110:6813","nonce":1405421093}]},"public_addr":"192.168.123.110:6809/1405421093","cluster_addr":"192.168.123.110:6811/1405421093","heartbeat_back_addr":"192.168.123.110:6815/1405421093","heartbeat_front_addr":"192.168.123.110:6813/1405421093","state":["exists","up"]},{"osd":6,"uuid":"6f33888e-59b3-4b64-acec-390be180037c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":38,"up_thru":39,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6816","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6817","nonce":1648093007}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6818","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6819","nonce":1648093007}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6822","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6823","nonce":1648093007}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6820","nonce":1648093007},{"type":"v1","addr":"192.168.123.110:6821","nonce":1648093007}]},"public_addr":"192.168.123.110:6817/1648093007","cluster_addr":"192.168.123.110:6819/1648093007","heartbeat_back_addr":"192.168.123.110:6823/1648093007","heartbeat_front_addr":"192.168.123.110:6821/1648093007","state":["exists","up"]},{"osd":7,"uuid":"6010e654-8935-42d2-9f1d-5fb78d19a935","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":43,"up_thru":44,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6824","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6825","nonce":303050948}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6826","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6827","nonce":303050948}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6830","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6831","nonce":303050948}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6828","nonce":303050948},{"type":"v1","addr":"192.168.123.110:6829","nonce":303050948}]},"public_addr":"192.168.123.110:6825/303050948","cluster_addr":"192.168.123.110:6827/303050948","heartbeat_back_addr":"192.168.123.110:6831/303050948","heartbeat_front_addr":"192.168.123.110:6829/303050948","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:10.341306+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:20.310352+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:28.227983+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:37.316803+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:45.441537+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:55:52.730521+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:56:01.814432+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T13:56:11.151529+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.108:6800/1786960600":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6801/1786960600":"2026-03-10T13:54:42.843895+0000","192.168.123.108:0/2425398216":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6800/2399132916":"2026-03-10T13:54:01.629885+0000","192.168.123.108:6801/2399132916":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/246681717":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/2203639542":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/3049100618":"2026-03-10T13:54:01.629885+0000","192.168.123.108:0/2992765302":"2026-03-10T13:54:42.843895+0000","192.168.123.108:0/489616954":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/3582363488":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/46790109":"2026-03-10T13:54:42.843895+0000","192.168.123.108:6800/677438560":"2026-03-10T13:53:49.429024+0000","192.168.123.108:6801/677438560":"2026-03-10T13:53:49.429024+0000","192.168.123.108:0/2681069985":"2026-03-10T13:53:49.429024+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T13:56:16.602 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph tell osd.0 flush_pg_stats 2026-03-09T13:56:16.603 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph tell osd.1 flush_pg_stats 2026-03-09T13:56:16.603 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph tell osd.2 flush_pg_stats 2026-03-09T13:56:16.603 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph tell osd.3 flush_pg_stats 2026-03-09T13:56:16.603 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph tell osd.4 flush_pg_stats 2026-03-09T13:56:16.603 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph tell osd.5 flush_pg_stats 2026-03-09T13:56:16.603 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph tell osd.6 flush_pg_stats 2026-03-09T13:56:16.603 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph tell osd.7 flush_pg_stats 2026-03-09T13:56:17.246 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:16 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1971362476' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T13:56:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:16 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1935088747' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T13:56:17.312 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:17.319 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:17.320 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:17.322 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:17.324 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:17.346 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:17.374 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:16 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/1971362476' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T13:56:17.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:16 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/1935088747' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T13:56:18.153 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:17 vm08 ceph-mon[47734]: pgmap v80: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:18.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:17 vm10 ceph-mon[51861]: pgmap v80: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:18.573 INFO:teuthology.orchestra.run.vm08.stdout:103079215113 2026-03-09T13:56:18.573 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd last-stat-seq osd.3 2026-03-09T13:56:18.701 INFO:teuthology.orchestra.run.vm08.stdout:73014444043 2026-03-09T13:56:18.701 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd last-stat-seq osd.2 2026-03-09T13:56:18.827 INFO:teuthology.orchestra.run.vm08.stdout:38654705679 2026-03-09T13:56:18.827 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd last-stat-seq osd.0 2026-03-09T13:56:18.828 INFO:teuthology.orchestra.run.vm08.stdout:141733920774 2026-03-09T13:56:18.828 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd last-stat-seq osd.5 2026-03-09T13:56:18.855 INFO:teuthology.orchestra.run.vm08.stdout:184683593731 2026-03-09T13:56:18.856 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd last-stat-seq osd.7 2026-03-09T13:56:18.936 INFO:teuthology.orchestra.run.vm08.stdout:163208757253 2026-03-09T13:56:18.936 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd last-stat-seq osd.6 2026-03-09T13:56:18.976 INFO:teuthology.orchestra.run.vm08.stdout:120259084296 2026-03-09T13:56:18.977 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd last-stat-seq osd.4 2026-03-09T13:56:18.994 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:19.053 INFO:teuthology.orchestra.run.vm08.stdout:55834574861 2026-03-09T13:56:19.053 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph osd last-stat-seq osd.1 2026-03-09T13:56:19.322 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:19.679 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:19.749 INFO:teuthology.orchestra.run.vm08.stdout:103079215114 2026-03-09T13:56:19.868 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:19.870 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:19.892 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:19.938 INFO:tasks.cephadm.ceph_manager.ceph:need seq 103079215113 got 103079215114 for osd.3 2026-03-09T13:56:19.938 DEBUG:teuthology.parallel:result is None 2026-03-09T13:56:19.941 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:20.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:19 vm08 ceph-mon[47734]: pgmap v81: 1 pgs: 1 active+recovering; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:20.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:19 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3781755417' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T13:56:20.335 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:20.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:19 vm10 ceph-mon[51861]: pgmap v81: 1 pgs: 1 active+recovering; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:20.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:19 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3781755417' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T13:56:20.535 INFO:teuthology.orchestra.run.vm08.stdout:73014444043 2026-03-09T13:56:20.778 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444043 got 73014444043 for osd.2 2026-03-09T13:56:20.778 DEBUG:teuthology.parallel:result is None 2026-03-09T13:56:21.006 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:20 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/738026150' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T13:56:21.006 INFO:teuthology.orchestra.run.vm08.stdout:163208757253 2026-03-09T13:56:21.083 INFO:teuthology.orchestra.run.vm08.stdout:120259084296 2026-03-09T13:56:21.122 INFO:teuthology.orchestra.run.vm08.stdout:38654705679 2026-03-09T13:56:21.204 INFO:tasks.cephadm.ceph_manager.ceph:need seq 120259084296 got 120259084296 for osd.4 2026-03-09T13:56:21.205 DEBUG:teuthology.parallel:result is None 2026-03-09T13:56:21.215 INFO:tasks.cephadm.ceph_manager.ceph:need seq 163208757253 got 163208757253 for osd.6 2026-03-09T13:56:21.215 DEBUG:teuthology.parallel:result is None 2026-03-09T13:56:21.319 INFO:tasks.cephadm.ceph_manager.ceph:need seq 38654705679 got 38654705679 for osd.0 2026-03-09T13:56:21.319 INFO:teuthology.orchestra.run.vm08.stdout:55834574861 2026-03-09T13:56:21.319 DEBUG:teuthology.parallel:result is None 2026-03-09T13:56:21.357 INFO:teuthology.orchestra.run.vm08.stdout:141733920775 2026-03-09T13:56:21.372 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574861 got 55834574861 for osd.1 2026-03-09T13:56:21.372 DEBUG:teuthology.parallel:result is None 2026-03-09T13:56:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:20 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/738026150' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T13:56:21.431 INFO:teuthology.orchestra.run.vm08.stdout:184683593731 2026-03-09T13:56:21.438 INFO:tasks.cephadm.ceph_manager.ceph:need seq 141733920774 got 141733920775 for osd.5 2026-03-09T13:56:21.438 DEBUG:teuthology.parallel:result is None 2026-03-09T13:56:21.477 INFO:tasks.cephadm.ceph_manager.ceph:need seq 184683593731 got 184683593731 for osd.7 2026-03-09T13:56:21.477 DEBUG:teuthology.parallel:result is None 2026-03-09T13:56:21.477 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-09T13:56:21.477 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph pg dump --format=json 2026-03-09T13:56:21.662 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:21.999 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:22.001 INFO:teuthology.orchestra.run.vm08.stderr:dumped all 2026-03-09T13:56:22.050 INFO:teuthology.orchestra.run.vm08.stdout:{"pg_ready":true,"pg_map":{"version":82,"stamp":"2026-03-09T13:56:20.920861+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":70160,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":5,"num_osds":8,"num_per_pool_osds":5,"num_per_pool_omap_osds":5,"kb":167739392,"kb_used":49088,"kb_used_data":5056,"kb_used_omap":0,"kb_used_meta":43968,"kb_avail":167690304,"statfs":{"total":171765137408,"available":171714871296,"internally_reserved":0,"allocated":5177344,"data_stored":3045500,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45023232},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"8.001007"},"pg_stats":[{"pgid":"1.0","version":"21'76","reported_seq":13,"reported_epoch":45,"state":"active+recovering","last_fresh":"2026-03-09T13:56:16.613495+0000","last_change":"2026-03-09T13:56:15.749719+0000","last_active":"2026-03-09T13:56:16.613495+0000","last_peered":"2026-03-09T13:56:16.613495+0000","last_clean":"2026-03-09T13:55:36.699644+0000","last_became_active":"2026-03-09T13:56:14.882397+0000","last_became_peered":"2026-03-09T13:56:14.882397+0000","last_unstale":"2026-03-09T13:56:16.613495+0000","last_undegraded":"2026-03-09T13:56:16.613495+0000","last_fullsized":"2026-03-09T13:56:16.613495+0000","mapping_epoch":44,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":40,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_clean_scrub_stamp":"2026-03-09T13:55:31.888202+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T00:26:09.984392+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":70160,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":["7","0","1","5","6"],"object_location_counts":[{"shards":"0,1,6,7","objects":1},{"shards":"0,6,7","objects":1}],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":70160,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1605632,"data_stored":1591360,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":43,"seq":184683593731,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60699999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.54100000000000004}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.542}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52800000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.94299999999999995}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47299999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.554}]}]},{"osd":6,"up_from":38,"seq":163208757253,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81899999999999995}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.90400000000000003}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67300000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.89100000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.1140000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.381}]}]},{"osd":1,"up_from":13,"seq":55834574861,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6656,"kb_used_data":824,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20960768,"statfs":{"total":21470642176,"available":21463826432,"internally_reserved":0,"allocated":843776,"data_stored":579328,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53400000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.66100000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70999999999999996}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58199999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52000000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]}]},{"osd":0,"up_from":9,"seq":38654705679,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6728,"kb_used_data":832,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960696,"statfs":{"total":21470642176,"available":21463752704,"internally_reserved":0,"allocated":851968,"data_stored":579887,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.875}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.54100000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.49099999999999999}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47199999999999998}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.48199999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83199999999999996}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60099999999999998}]}]},{"osd":2,"up_from":17,"seq":73014444044,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6272,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961152,"statfs":{"total":21470642176,"available":21464219648,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86199999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71599999999999997}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.76500000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82099999999999995}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.745}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83899999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]}]},{"osd":3,"up_from":24,"seq":103079215114,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5756,"kb_used_data":436,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961668,"statfs":{"total":21470642176,"available":21464748032,"internally_reserved":0,"allocated":446464,"data_stored":181803,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53900000000000003}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.55400000000000005}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51000000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58799999999999997}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53200000000000003}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63400000000000001}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.097}]}]},{"osd":4,"up_from":28,"seq":120259084296,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":181244,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58199999999999996}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.46999999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.38500000000000001}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51800000000000002}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.33200000000000002}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.68200000000000005}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.89200000000000002}]}]},{"osd":5,"up_from":33,"seq":141733920775,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":5760,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961664,"statfs":{"total":21470642176,"available":21464743936,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53300000000000003}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51400000000000001}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79600000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.77700000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57299999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59599999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83499999999999996}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T13:56:22.051 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph pg dump --format=json 2026-03-09T13:56:22.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:21 vm08 ceph-mon[47734]: pgmap v82: 1 pgs: 1 active+recovering; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:22.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:21 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2920229567' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T13:56:22.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:21 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/825476413' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T13:56:22.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:21 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/202651108' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T13:56:22.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:21 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1561138761' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T13:56:22.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:21 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3406694561' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T13:56:22.051 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:21 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/3804889693' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-09T13:56:22.206 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:21 vm10 ceph-mon[51861]: pgmap v82: 1 pgs: 1 active+recovering; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:21 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2920229567' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T13:56:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:21 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/825476413' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T13:56:22.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:21 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/202651108' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T13:56:22.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:21 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/1561138761' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T13:56:22.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:21 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3406694561' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T13:56:22.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:21 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/3804889693' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-09T13:56:22.534 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:22.538 INFO:teuthology.orchestra.run.vm08.stderr:dumped all 2026-03-09T13:56:22.580 INFO:teuthology.orchestra.run.vm08.stdout:{"pg_ready":true,"pg_map":{"version":82,"stamp":"2026-03-09T13:56:20.920861+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":70160,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":5,"num_osds":8,"num_per_pool_osds":5,"num_per_pool_omap_osds":5,"kb":167739392,"kb_used":49088,"kb_used_data":5056,"kb_used_omap":0,"kb_used_meta":43968,"kb_avail":167690304,"statfs":{"total":171765137408,"available":171714871296,"internally_reserved":0,"allocated":5177344,"data_stored":3045500,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45023232},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"8.001007"},"pg_stats":[{"pgid":"1.0","version":"21'76","reported_seq":13,"reported_epoch":45,"state":"active+recovering","last_fresh":"2026-03-09T13:56:16.613495+0000","last_change":"2026-03-09T13:56:15.749719+0000","last_active":"2026-03-09T13:56:16.613495+0000","last_peered":"2026-03-09T13:56:16.613495+0000","last_clean":"2026-03-09T13:55:36.699644+0000","last_became_active":"2026-03-09T13:56:14.882397+0000","last_became_peered":"2026-03-09T13:56:14.882397+0000","last_unstale":"2026-03-09T13:56:16.613495+0000","last_undegraded":"2026-03-09T13:56:16.613495+0000","last_fullsized":"2026-03-09T13:56:16.613495+0000","mapping_epoch":44,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":40,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_clean_scrub_stamp":"2026-03-09T13:55:31.888202+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T00:26:09.984392+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":70160,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":["7","0","1","5","6"],"object_location_counts":[{"shards":"0,1,6,7","objects":1},{"shards":"0,6,7","objects":1}],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":70160,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1605632,"data_stored":1591360,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":43,"seq":184683593731,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60699999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.54100000000000004}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.542}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52800000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.94299999999999995}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47299999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.554}]}]},{"osd":6,"up_from":38,"seq":163208757253,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81899999999999995}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.90400000000000003}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67300000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.89100000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.1140000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.381}]}]},{"osd":1,"up_from":13,"seq":55834574861,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6656,"kb_used_data":824,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20960768,"statfs":{"total":21470642176,"available":21463826432,"internally_reserved":0,"allocated":843776,"data_stored":579328,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53400000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.66100000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70999999999999996}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58199999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52000000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]}]},{"osd":0,"up_from":9,"seq":38654705679,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6728,"kb_used_data":832,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960696,"statfs":{"total":21470642176,"available":21463752704,"internally_reserved":0,"allocated":851968,"data_stored":579887,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.875}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.54100000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.49099999999999999}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47199999999999998}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.48199999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83199999999999996}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60099999999999998}]}]},{"osd":2,"up_from":17,"seq":73014444044,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6272,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961152,"statfs":{"total":21470642176,"available":21464219648,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86199999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71599999999999997}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.76500000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82099999999999995}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.745}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83899999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]}]},{"osd":3,"up_from":24,"seq":103079215114,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5756,"kb_used_data":436,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961668,"statfs":{"total":21470642176,"available":21464748032,"internally_reserved":0,"allocated":446464,"data_stored":181803,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53900000000000003}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.55400000000000005}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51000000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58799999999999997}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53200000000000003}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63400000000000001}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.097}]}]},{"osd":4,"up_from":28,"seq":120259084296,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":181244,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58199999999999996}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.46999999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.38500000000000001}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51800000000000002}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.33200000000000002}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.68200000000000005}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.89200000000000002}]}]},{"osd":5,"up_from":33,"seq":141733920775,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":5760,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961664,"statfs":{"total":21470642176,"available":21464743936,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53300000000000003}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51400000000000001}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79600000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.77700000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57299999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59599999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83499999999999996}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T13:56:22.581 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph pg dump --format=json 2026-03-09T13:56:22.731 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:23.054 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:23.057 INFO:teuthology.orchestra.run.vm08.stderr:dumped all 2026-03-09T13:56:23.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:22 vm08 ceph-mon[47734]: from='client.24303 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:23.095 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:22 vm08 ceph-mon[47734]: from='client.14496 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:23.124 INFO:teuthology.orchestra.run.vm08.stdout:{"pg_ready":true,"pg_map":{"version":83,"stamp":"2026-03-09T13:56:22.921137+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":5,"num_osds":8,"num_per_pool_osds":5,"num_per_pool_omap_osds":5,"kb":167739392,"kb_used":48696,"kb_used_data":4664,"kb_used_omap":0,"kb_used_meta":43968,"kb_avail":167690696,"statfs":{"total":171765137408,"available":171715272704,"internally_reserved":0,"allocated":4775936,"data_stored":2647660,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45023232},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":327680,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"9.046732"},"pg_stats":[{"pgid":"1.0","version":"21'76","reported_seq":17,"reported_epoch":45,"state":"active+clean","last_fresh":"2026-03-09T13:56:20.673045+0000","last_change":"2026-03-09T13:56:20.673045+0000","last_active":"2026-03-09T13:56:20.673045+0000","last_peered":"2026-03-09T13:56:20.673045+0000","last_clean":"2026-03-09T13:56:20.673045+0000","last_became_active":"2026-03-09T13:56:14.882397+0000","last_became_peered":"2026-03-09T13:56:14.882397+0000","last_unstale":"2026-03-09T13:56:20.673045+0000","last_undegraded":"2026-03-09T13:56:20.673045+0000","last_fullsized":"2026-03-09T13:56:20.673045+0000","mapping_epoch":44,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":45,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_clean_scrub_stamp":"2026-03-09T13:55:31.888202+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T00:26:09.984392+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1204224,"data_stored":1193520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":43,"seq":184683593732,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72499999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59999999999999998}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86399999999999999}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59199999999999997}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81000000000000005}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57399999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.92500000000000004}]}]},{"osd":6,"up_from":38,"seq":163208757253,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81899999999999995}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.90400000000000003}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67300000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.89100000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.1140000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.381}]}]},{"osd":1,"up_from":13,"seq":55834574862,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6264,"kb_used_data":432,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961160,"statfs":{"total":21470642176,"available":21464227840,"internally_reserved":0,"allocated":442368,"data_stored":181488,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53400000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.66100000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70999999999999996}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58199999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52000000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]}]},{"osd":0,"up_from":9,"seq":38654705680,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6728,"kb_used_data":832,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960696,"statfs":{"total":21470642176,"available":21463752704,"internally_reserved":0,"allocated":851968,"data_stored":579887,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81699999999999995}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.34100000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.49099999999999999}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.92600000000000005}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47799999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52500000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.93300000000000005}]}]},{"osd":2,"up_from":17,"seq":73014444044,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6272,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961152,"statfs":{"total":21470642176,"available":21464219648,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86199999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71599999999999997}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.76500000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82099999999999995}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.745}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83899999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]}]},{"osd":3,"up_from":24,"seq":103079215114,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5756,"kb_used_data":436,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961668,"statfs":{"total":21470642176,"available":21464748032,"internally_reserved":0,"allocated":446464,"data_stored":181803,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53900000000000003}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.55400000000000005}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51000000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58799999999999997}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53200000000000003}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63400000000000001}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.097}]}]},{"osd":4,"up_from":28,"seq":120259084297,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":181244,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47699999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":12.718999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.55400000000000005}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":12.637}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.49299999999999999}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51300000000000001}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.755}]}]},{"osd":5,"up_from":33,"seq":141733920775,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":5760,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961664,"statfs":{"total":21470642176,"available":21464743936,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53300000000000003}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51400000000000001}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79600000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.77700000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57299999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59599999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83499999999999996}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T13:56:23.124 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph status --format=json 2026-03-09T13:56:23.271 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:22 vm10 ceph-mon[51861]: from='client.24303 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:22 vm10 ceph-mon[51861]: from='client.14496 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:23.645 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:23.645 INFO:teuthology.orchestra.run.vm08.stdout:{"fsid":"4a3e67a6-1bbf-11f1-aca1-21c5f680cc64","health":{"status":"HEALTH_OK","checks":{},"mutes":[]},"election_epoch":10,"quorum":[0,1],"quorum_names":["vm08","vm10"],"quorum_age":89,"monmap":{"epoch":2,"min_mon_release_name":"quincy","num_mons":2},"osdmap":{"epoch":45,"num_osds":8,"num_up_osds":8,"osd_up_since":1773064572,"num_in_osds":8,"osd_in_since":1773064563,"num_remapped_pgs":0},"pgmap":{"pgs_by_state":[{"state_name":"active+clean","count":1}],"num_pgs":1,"num_pools":1,"num_objects":2,"data_bytes":459280,"bytes_used":49864704,"bytes_avail":171715272704,"bytes_total":171765137408,"recovering_objects_per_sec":0,"recovering_bytes_per_sec":36220,"recovering_keys_per_sec":0,"num_objects_recovered":1,"num_bytes_recovered":327680,"num_keys_recovered":0},"fsmap":{"epoch":1,"by_rank":[],"up:standby":0},"mgrmap":{"available":true,"num_standbys":1,"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"services":{"dashboard":"https://192.168.123.108:8443/","prometheus":"http://192.168.123.108:9283/"}},"servicemap":{"epoch":6,"modified":"2026-03-09T13:56:22.921200+0000","services":{"osd":{"daemons":{"summary":"","2":{"start_epoch":0,"start_stamp":"0.000000","gid":0,"addr":"(unrecognized address family 0)/0","metadata":{},"task_status":{}},"3":{"start_epoch":0,"start_stamp":"0.000000","gid":0,"addr":"(unrecognized address family 0)/0","metadata":{},"task_status":{}},"4":{"start_epoch":0,"start_stamp":"0.000000","gid":0,"addr":"(unrecognized address family 0)/0","metadata":{},"task_status":{}},"5":{"start_epoch":0,"start_stamp":"0.000000","gid":0,"addr":"(unrecognized address family 0)/0","metadata":{},"task_status":{}},"6":{"start_epoch":0,"start_stamp":"0.000000","gid":0,"addr":"(unrecognized address family 0)/0","metadata":{},"task_status":{}},"7":{"start_epoch":0,"start_stamp":"0.000000","gid":0,"addr":"(unrecognized address family 0)/0","metadata":{},"task_status":{}}}}}},"progress_events":{}} 2026-03-09T13:56:23.693 INFO:tasks.cephadm.ceph_manager.ceph:making progress, resetting timeout 2026-03-09T13:56:23.693 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph pg dump --format=json 2026-03-09T13:56:23.851 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:24.192 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:24.193 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:23 vm08 ceph-mon[47734]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 35 KiB/s, 0 objects/s recovering 2026-03-09T13:56:24.193 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:23 vm08 ceph-mon[47734]: from='client.14500 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:24.193 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:23 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/1796677886' entity='client.admin' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T13:56:24.195 INFO:teuthology.orchestra.run.vm08.stderr:dumped all 2026-03-09T13:56:24.261 INFO:teuthology.orchestra.run.vm08.stdout:{"pg_ready":true,"pg_map":{"version":83,"stamp":"2026-03-09T13:56:22.921137+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":5,"num_osds":8,"num_per_pool_osds":5,"num_per_pool_omap_osds":5,"kb":167739392,"kb_used":48696,"kb_used_data":4664,"kb_used_omap":0,"kb_used_meta":43968,"kb_avail":167690696,"statfs":{"total":171765137408,"available":171715272704,"internally_reserved":0,"allocated":4775936,"data_stored":2647660,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45023232},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":327680,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"9.046732"},"pg_stats":[{"pgid":"1.0","version":"21'76","reported_seq":17,"reported_epoch":45,"state":"active+clean","last_fresh":"2026-03-09T13:56:20.673045+0000","last_change":"2026-03-09T13:56:20.673045+0000","last_active":"2026-03-09T13:56:20.673045+0000","last_peered":"2026-03-09T13:56:20.673045+0000","last_clean":"2026-03-09T13:56:20.673045+0000","last_became_active":"2026-03-09T13:56:14.882397+0000","last_became_peered":"2026-03-09T13:56:14.882397+0000","last_unstale":"2026-03-09T13:56:20.673045+0000","last_undegraded":"2026-03-09T13:56:20.673045+0000","last_fullsized":"2026-03-09T13:56:20.673045+0000","mapping_epoch":44,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":45,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_clean_scrub_stamp":"2026-03-09T13:55:31.888202+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T00:26:09.984392+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1204224,"data_stored":1193520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":43,"seq":184683593732,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72499999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59999999999999998}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86399999999999999}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59199999999999997}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81000000000000005}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57399999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.92500000000000004}]}]},{"osd":6,"up_from":38,"seq":163208757253,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81899999999999995}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.90400000000000003}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67300000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.89100000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.1140000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.381}]}]},{"osd":1,"up_from":13,"seq":55834574862,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6264,"kb_used_data":432,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961160,"statfs":{"total":21470642176,"available":21464227840,"internally_reserved":0,"allocated":442368,"data_stored":181488,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53400000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.66100000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70999999999999996}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58199999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52000000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]}]},{"osd":0,"up_from":9,"seq":38654705680,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6728,"kb_used_data":832,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960696,"statfs":{"total":21470642176,"available":21463752704,"internally_reserved":0,"allocated":851968,"data_stored":579887,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81699999999999995}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.34100000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.49099999999999999}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.92600000000000005}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47799999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52500000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.93300000000000005}]}]},{"osd":2,"up_from":17,"seq":73014444044,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6272,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961152,"statfs":{"total":21470642176,"available":21464219648,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86199999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71599999999999997}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.76500000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82099999999999995}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.745}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83899999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]}]},{"osd":3,"up_from":24,"seq":103079215114,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5756,"kb_used_data":436,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961668,"statfs":{"total":21470642176,"available":21464748032,"internally_reserved":0,"allocated":446464,"data_stored":181803,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53900000000000003}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.55400000000000005}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51000000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58799999999999997}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53200000000000003}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63400000000000001}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.097}]}]},{"osd":4,"up_from":28,"seq":120259084297,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":181244,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47699999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":12.718999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.55400000000000005}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":12.637}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.49299999999999999}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51300000000000001}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.755}]}]},{"osd":5,"up_from":33,"seq":141733920775,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":5760,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961664,"statfs":{"total":21470642176,"available":21464743936,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53300000000000003}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.51400000000000001}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79600000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.77700000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57299999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59599999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83499999999999996}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T13:56:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:23 vm10 ceph-mon[51861]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 35 KiB/s, 0 objects/s recovering 2026-03-09T13:56:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:23 vm10 ceph-mon[51861]: from='client.14500 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:23 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/1796677886' entity='client.admin' cmd=[{"prefix": "status", "format": "json"}]: dispatch 2026-03-09T13:56:25.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:25 vm08 ceph-mon[47734]: from='client.14508 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:25 vm10 ceph-mon[51861]: from='client.14508 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:26.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:26 vm08 ceph-mon[47734]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s, 0 objects/s recovering 2026-03-09T13:56:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:26 vm10 ceph-mon[51861]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s, 0 objects/s recovering 2026-03-09T13:56:27.262 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph pg dump --format=json 2026-03-09T13:56:27.409 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:27.747 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:27.749 INFO:teuthology.orchestra.run.vm08.stderr:dumped all 2026-03-09T13:56:27.811 INFO:teuthology.orchestra.run.vm08.stdout:{"pg_ready":true,"pg_map":{"version":85,"stamp":"2026-03-09T13:56:26.921747+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":4,"num_osds":8,"num_per_pool_osds":4,"num_per_pool_omap_osds":4,"kb":167739392,"kb_used":48696,"kb_used_data":4664,"kb_used_omap":0,"kb_used_meta":43968,"kb_avail":167690696,"statfs":{"total":171765137408,"available":171715272704,"internally_reserved":0,"allocated":4775936,"data_stored":2647660,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45023232},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":327680,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"12.001637"},"pg_stats":[{"pgid":"1.0","version":"21'76","reported_seq":17,"reported_epoch":45,"state":"active+clean","last_fresh":"2026-03-09T13:56:20.673045+0000","last_change":"2026-03-09T13:56:20.673045+0000","last_active":"2026-03-09T13:56:20.673045+0000","last_peered":"2026-03-09T13:56:20.673045+0000","last_clean":"2026-03-09T13:56:20.673045+0000","last_became_active":"2026-03-09T13:56:14.882397+0000","last_became_peered":"2026-03-09T13:56:14.882397+0000","last_unstale":"2026-03-09T13:56:20.673045+0000","last_undegraded":"2026-03-09T13:56:20.673045+0000","last_fullsized":"2026-03-09T13:56:20.673045+0000","mapping_epoch":44,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":45,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T13:55:31.888202+0000","last_clean_scrub_stamp":"2026-03-09T13:55:31.888202+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T00:26:09.984392+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1204224,"data_stored":1193520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":43,"seq":184683593732,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72499999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59999999999999998}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86399999999999999}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59199999999999997}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.81000000000000005}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57399999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.92500000000000004}]}]},{"osd":6,"up_from":38,"seq":163208757254,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6084,"kb_used_data":828,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961340,"statfs":{"total":21470642176,"available":21464412160,"internally_reserved":0,"allocated":847872,"data_stored":579572,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.73499999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.68999999999999995}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79800000000000004}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65700000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.55300000000000005}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52400000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.873}]}]},{"osd":1,"up_from":13,"seq":55834574862,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6264,"kb_used_data":432,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961160,"statfs":{"total":21470642176,"available":21464227840,"internally_reserved":0,"allocated":442368,"data_stored":181488,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53400000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.66100000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70999999999999996}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58199999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52000000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72899999999999998}]}]},{"osd":0,"up_from":9,"seq":38654705681,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6728,"kb_used_data":832,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960696,"statfs":{"total":21470642176,"available":21463752704,"internally_reserved":0,"allocated":851968,"data_stored":579887,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.625}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.40799999999999997}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57499999999999996}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70199999999999996}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.56499999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59799999999999998}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69399999999999995}]}]},{"osd":2,"up_from":17,"seq":73014444045,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6272,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961152,"statfs":{"total":21470642176,"available":21464219648,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86199999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71599999999999997}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.76500000000000001}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82099999999999995}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.745}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.83899999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59399999999999997}]}]},{"osd":3,"up_from":24,"seq":103079215115,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5756,"kb_used_data":436,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961668,"statfs":{"total":21470642176,"available":21464748032,"internally_reserved":0,"allocated":446464,"data_stored":181803,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.30499999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.56999999999999995}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.54600000000000004}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60499999999999998}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.56100000000000005}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67100000000000004}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79700000000000004}]}]},{"osd":4,"up_from":28,"seq":120259084298,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":181244,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.42699999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.47599999999999998}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.61299999999999999}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71199999999999997}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.41399999999999998}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69599999999999995}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.71899999999999997}]}]},{"osd":5,"up_from":33,"seq":141733920776,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5760,"kb_used_data":440,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961664,"statfs":{"total":21470642176,"available":21464743936,"internally_reserved":0,"allocated":450560,"data_stored":182047,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.61899999999999999}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.628}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60299999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.69199999999999995}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70899999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.751}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65100000000000002}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T13:56:27.811 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-09T13:56:27.811 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-09T13:56:27.811 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-09T13:56:27.811 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph health --format=json 2026-03-09T13:56:27.962 INFO:teuthology.orchestra.run.vm08.stderr:Inferring config /var/lib/ceph/4a3e67a6-1bbf-11f1-aca1-21c5f680cc64/mon.vm08/config 2026-03-09T13:56:28.041 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:28 vm08 ceph-mon[47734]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T13:56:28.313 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:28.313 INFO:teuthology.orchestra.run.vm08.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-09T13:56:28.359 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-09T13:56:28.359 INFO:tasks.cephadm:Setup complete, yielding 2026-03-09T13:56:28.359 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T13:56:28.361 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm08.local 2026-03-09T13:56:28.361 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'ceph orch status' 2026-03-09T13:56:28.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:28 vm10 ceph-mon[51861]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T13:56:28.818 INFO:teuthology.orchestra.run.vm08.stdout:Backend: cephadm 2026-03-09T13:56:28.818 INFO:teuthology.orchestra.run.vm08.stdout:Available: Yes 2026-03-09T13:56:28.818 INFO:teuthology.orchestra.run.vm08.stdout:Paused: No 2026-03-09T13:56:28.889 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'ceph orch ps' 2026-03-09T13:56:29.091 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:29 vm08 ceph-mon[47734]: from='client.14512 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:29.091 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:29 vm08 ceph-mon[47734]: from='client.? 192.168.123.108:0/2118847823' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:alertmanager.vm08 vm08 *:9093,9094 running (81s) 53s ago 2m 17.9M - ba2b418f427c e3a11ed0b6c4 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:crash.vm08 vm08 running (2m) 53s ago 2m 6945k - 17.2.0 e1d6a67b021e 991839d3aa3a 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:crash.vm10 vm10 running (102s) 19s ago 102s 7105k - 17.2.0 e1d6a67b021e dc7bc9a567ed 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:grafana.vm08 vm08 *:3000 running (79s) 53s ago 114s 40.5M - 8.3.5 dad864ee21e9 dd22b0a0ca57 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:mgr.vm08.siqces vm08 *:9283 running (2m) 53s ago 2m 456M - 17.2.0 e1d6a67b021e 9ac551dcfa2c 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:mgr.vm10.euefis vm10 *:8443,9283 running (101s) 19s ago 101s 420M - 17.2.0 e1d6a67b021e 48b8b2d0c40d 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:mon.vm08 vm08 running (2m) 53s ago 2m 41.6M 2048M 17.2.0 e1d6a67b021e b2deb9231198 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:mon.vm10 vm10 running (100s) 19s ago 100s 39.7M 2048M 17.2.0 e1d6a67b021e 93f2f81fb24e 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:node-exporter.vm08 vm08 *:9100 running (111s) 53s ago 111s 18.2M - 1dbe0e931976 fd80ded96dbd 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:node-exporter.vm10 vm10 *:9100 running (97s) 19s ago 97s 19.4M - 1dbe0e931976 bd244dc15059 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:osd.0 vm08 running (81s) 53s ago 81s 33.7M 4096M 17.2.0 e1d6a67b021e ff15d5c3a685 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:osd.1 vm08 running (72s) 53s ago 71s 35.0M 4096M 17.2.0 e1d6a67b021e 1c5c066a8f7e 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:osd.2 vm08 running (63s) 53s ago 63s 30.0M 4096M 17.2.0 e1d6a67b021e ef8af621dacd 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:osd.3 vm08 running (55s) 53s ago 55s 12.6M 4096M 17.2.0 e1d6a67b021e 6c250cb61804 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:osd.4 vm10 running (46s) 19s ago 46s 32.3M 4096M 17.2.0 e1d6a67b021e e56d41afc6d6 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:osd.5 vm10 running (39s) 19s ago 39s 32.8M 4096M 17.2.0 e1d6a67b021e b7dc3d7c4cd7 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:osd.6 vm10 running (30s) 19s ago 30s 30.8M 4096M 17.2.0 e1d6a67b021e d2f5d91f6ee2 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:osd.7 vm10 running (21s) 19s ago 20s 11.2M 4096M 17.2.0 e1d6a67b021e 71e224c5da5c 2026-03-09T13:56:29.376 INFO:teuthology.orchestra.run.vm08.stdout:prometheus.vm08 vm08 *:9095 running (89s) 53s ago 89s 36.7M - 514e6a882f6e 7c85a4da167d 2026-03-09T13:56:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:29 vm10 ceph-mon[51861]: from='client.14512 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:29 vm10 ceph-mon[51861]: from='client.? 192.168.123.108:0/2118847823' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T13:56:29.501 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'ceph orch ls' 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:alertmanager ?:9093,9094 1/1 53s ago 2m count:1 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:crash 2/2 53s ago 2m * 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:grafana ?:3000 1/1 53s ago 2m count:1 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:mgr 2/2 53s ago 2m count:2 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:mon 2/2 53s ago 2m vm08:192.168.123.108=vm08;vm10:192.168.123.110=vm10;count:2 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:node-exporter ?:9100 2/2 53s ago 2m * 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:osd 8 53s ago - 2026-03-09T13:56:29.976 INFO:teuthology.orchestra.run.vm08.stdout:prometheus ?:9095 1/1 53s ago 2m count:1 2026-03-09T13:56:30.066 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'ceph orch host ls' 2026-03-09T13:56:30.241 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:30 vm08 ceph-mon[47734]: from='client.14520 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:30.241 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:30 vm08 ceph-mon[47734]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T13:56:30.241 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:30 vm08 ceph-mon[47734]: from='client.14524 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:30.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:30 vm10 ceph-mon[51861]: from='client.14520 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:30.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:30 vm10 ceph-mon[51861]: pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T13:56:30.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:30 vm10 ceph-mon[51861]: from='client.14524 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:30.593 INFO:teuthology.orchestra.run.vm08.stdout:HOST ADDR LABELS STATUS 2026-03-09T13:56:30.593 INFO:teuthology.orchestra.run.vm08.stdout:vm08 192.168.123.108 2026-03-09T13:56:30.593 INFO:teuthology.orchestra.run.vm08.stdout:vm10 192.168.123.110 2026-03-09T13:56:30.593 INFO:teuthology.orchestra.run.vm08.stdout:2 hosts in cluster 2026-03-09T13:56:30.651 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'ceph orch device ls' 2026-03-09T13:56:31.180 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:31 vm08 ceph-mon[47734]: from='client.14528 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:31.180 INFO:teuthology.orchestra.run.vm08.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REJECT REASONS 2026-03-09T13:56:31.180 INFO:teuthology.orchestra.run.vm08.stdout:vm08 /dev/vdb hdd DWNBRSTVMM08001 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T13:56:31.180 INFO:teuthology.orchestra.run.vm08.stdout:vm08 /dev/vdc hdd DWNBRSTVMM08002 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T13:56:31.180 INFO:teuthology.orchestra.run.vm08.stdout:vm08 /dev/vdd hdd DWNBRSTVMM08003 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T13:56:31.180 INFO:teuthology.orchestra.run.vm08.stdout:vm08 /dev/vde hdd DWNBRSTVMM08004 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T13:56:31.180 INFO:teuthology.orchestra.run.vm08.stdout:vm10 /dev/vdb hdd DWNBRSTVMM10001 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T13:56:31.181 INFO:teuthology.orchestra.run.vm08.stdout:vm10 /dev/vdc hdd DWNBRSTVMM10002 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T13:56:31.181 INFO:teuthology.orchestra.run.vm08.stdout:vm10 /dev/vdd hdd DWNBRSTVMM10003 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T13:56:31.181 INFO:teuthology.orchestra.run.vm08.stdout:vm10 /dev/vde hdd DWNBRSTVMM10004 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T13:56:31.302 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-09T13:56:31.306 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm08.local 2026-03-09T13:56:31.306 DEBUG:teuthology.orchestra.run.vm08:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-09T13:56:31.337 INFO:teuthology.orchestra.run.vm08.stderr:+ systemctl stop nfs-server 2026-03-09T13:56:31.346 INFO:tasks.vip:Running commands on role host.b host ubuntu@vm10.local 2026-03-09T13:56:31.347 DEBUG:teuthology.orchestra.run.vm10:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-09T13:56:31.373 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:31 vm10 ceph-mon[51861]: from='client.14528 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:31.376 INFO:teuthology.orchestra.run.vm10.stderr:+ systemctl stop nfs-server 2026-03-09T13:56:31.383 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T13:56:31.385 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm08.local 2026-03-09T13:56:31.385 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'ceph fs volume create foofs' 2026-03-09T13:56:32.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:32 vm10 ceph-mon[51861]: from='client.14532 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:32.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:32 vm10 ceph-mon[51861]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T13:56:32.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:32 vm10 ceph-mon[51861]: from='client.14536 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:32.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:32 vm08 ceph-mon[47734]: from='client.14532 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:32.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:32 vm08 ceph-mon[47734]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T13:56:32.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:32 vm08 ceph-mon[47734]: from='client.14536 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:33.323 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-09T13:56:33.325 INFO:tasks.cephadm:Waiting for ceph service mds.foofs to start (timeout 300)... 2026-03-09T13:56:33.326 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch ls -f json 2026-03-09T13:56:33.394 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:33 vm08 ceph-mon[47734]: from='client.14540 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "foofs", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:33.394 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:33 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]: dispatch 2026-03-09T13:56:33.394 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:33 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]': finished 2026-03-09T13:56:33.394 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:33 vm08 ceph-mon[47734]: osdmap e46: 8 total, 8 up, 8 in 2026-03-09T13:56:33.394 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:33 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]: dispatch 2026-03-09T13:56:33.394 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:33 vm08 ceph-4a3e67a6-1bbf-11f1-aca1-21c5f680cc64-mon-vm08[47730]: 2026-03-09T13:56:33.208+0000 7fe0ca485700 -1 log_channel(cluster) log [ERR] : Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-09T13:56:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:33 vm10 ceph-mon[51861]: from='client.14540 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "foofs", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:33 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]: dispatch 2026-03-09T13:56:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:33 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]': finished 2026-03-09T13:56:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:33 vm10 ceph-mon[51861]: osdmap e46: 8 total, 8 up, 8 in 2026-03-09T13:56:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:33 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]: dispatch 2026-03-09T13:56:33.941 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:33.941 INFO:teuthology.orchestra.run.vm08.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-09T13:53:55.374593Z", "last_refresh": "2026-03-09T13:55:36.364116Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-09T13:53:53.861815Z", "last_refresh": "2026-03-09T13:55:36.364146Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-09T13:53:54.676006Z", "last_refresh": "2026-03-09T13:55:36.364200Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-09T13:56:33.247282Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-09T13:56:33.240629Z", "running": 0, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-09T13:53:53.550627Z", "last_refresh": "2026-03-09T13:55:36.364076Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm08:192.168.123.108=vm08", "vm10:192.168.123.110=vm10"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-09T13:54:10.434622Z", "last_refresh": "2026-03-09T13:55:36.363962Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-09T13:53:55.049639Z", "last_refresh": "2026-03-09T13:55:36.364262Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-09T13:55:36.364318Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-09T13:53:54.164175Z", "last_refresh": "2026-03-09T13:55:36.364291Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-09T13:56:34.191 INFO:tasks.cephadm:mds.foofs has 0/2 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: pgmap v89: 33 pgs: 12 creating+peering, 20 unknown, 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]': finished 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: osdmap e47: 8 total, 8 up, 8 in 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]: dispatch 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: osdmap e48: 8 total, 8 up, 8 in 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]': finished 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: fsmap foofs:0 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: Saving service mds.foofs spec with placement count:2 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm08.jgklas", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm08.jgklas", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:34.441 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:34 vm10 ceph-mon[51861]: Deploying daemon mds.foofs.vm08.jgklas on vm08 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: pgmap v89: 33 pgs: 12 creating+peering, 20 unknown, 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]': finished 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: osdmap e47: 8 total, 8 up, 8 in 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]: dispatch 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: osdmap e48: 8 total, 8 up, 8 in 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]': finished 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: fsmap foofs:0 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: Saving service mds.foofs spec with placement count:2 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm08.jgklas", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm08.jgklas", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:34 vm08 ceph-mon[47734]: Deploying daemon mds.foofs.vm08.jgklas on vm08 2026-03-09T13:56:35.193 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch ls -f json 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: from='client.14544 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: osdmap e49: 8 total, 8 up, 8 in 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm10.uuhfwt", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm10.uuhfwt", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: Deploying daemon mds.foofs.vm10.uuhfwt on vm10 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: daemon mds.foofs.vm08.jgklas assigned to filesystem foofs as rank 0 (now has 1 ranks) 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: Cluster is now healthy 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: mds.? [v2:192.168.123.108:6834/3315308194,v1:192.168.123.108:6835/3315308194] up:boot 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: fsmap foofs:1 {0=foofs.vm08.jgklas=up:creating} 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mds metadata", "who": "foofs.vm08.jgklas"}]: dispatch 2026-03-09T13:56:35.334 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:35 vm10 ceph-mon[51861]: daemon mds.foofs.vm08.jgklas is now active in filesystem foofs as rank 0 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: from='client.14544 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: osdmap e49: 8 total, 8 up, 8 in 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm10.uuhfwt", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm10.uuhfwt", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: Deploying daemon mds.foofs.vm10.uuhfwt on vm10 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: daemon mds.foofs.vm08.jgklas assigned to filesystem foofs as rank 0 (now has 1 ranks) 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: Cluster is now healthy 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: mds.? [v2:192.168.123.108:6834/3315308194,v1:192.168.123.108:6835/3315308194] up:boot 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: fsmap foofs:1 {0=foofs.vm08.jgklas=up:creating} 2026-03-09T13:56:35.486 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mds metadata", "who": "foofs.vm08.jgklas"}]: dispatch 2026-03-09T13:56:35.487 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:35 vm08 ceph-mon[47734]: daemon mds.foofs.vm08.jgklas is now active in filesystem foofs as rank 0 2026-03-09T13:56:35.832 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:35.832 INFO:teuthology.orchestra.run.vm08.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-09T13:53:55.374593Z", "last_refresh": "2026-03-09T13:55:36.364116Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-09T13:53:53.861815Z", "last_refresh": "2026-03-09T13:55:36.364146Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-09T13:53:54.676006Z", "last_refresh": "2026-03-09T13:55:36.364200Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-09T13:56:33.247282Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-09T13:56:33.240629Z", "running": 0, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-09T13:53:53.550627Z", "last_refresh": "2026-03-09T13:55:36.364076Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm08:192.168.123.108=vm08", "vm10:192.168.123.110=vm10"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-09T13:54:10.434622Z", "last_refresh": "2026-03-09T13:55:36.363962Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-09T13:53:55.049639Z", "last_refresh": "2026-03-09T13:55:36.364262Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-09T13:55:36.364318Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-09T13:53:54.164175Z", "last_refresh": "2026-03-09T13:55:36.364291Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-09T13:56:35.921 INFO:tasks.cephadm:mds.foofs has 0/2 2026-03-09T13:56:36.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: pgmap v93: 65 pgs: 20 creating+peering, 22 unknown, 23 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:36.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: osdmap e50: 8 total, 8 up, 8 in 2026-03-09T13:56:36.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:36.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:36.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:36.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:36.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: mds.? [v2:192.168.123.108:6834/3315308194,v1:192.168.123.108:6835/3315308194] up:active 2026-03-09T13:56:36.308 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: mds.? [v2:192.168.123.110:6832/748767188,v1:192.168.123.110:6833/748767188] up:boot 2026-03-09T13:56:36.308 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: fsmap foofs:1 {0=foofs.vm08.jgklas=up:active} 1 up:standby 2026-03-09T13:56:36.308 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mds metadata", "who": "foofs.vm10.uuhfwt"}]: dispatch 2026-03-09T13:56:36.308 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: fsmap foofs:1 {0=foofs.vm08.jgklas=up:active} 1 up:standby 2026-03-09T13:56:36.308 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:36 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: pgmap v93: 65 pgs: 20 creating+peering, 22 unknown, 23 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: osdmap e50: 8 total, 8 up, 8 in 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: mds.? [v2:192.168.123.108:6834/3315308194,v1:192.168.123.108:6835/3315308194] up:active 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: mds.? [v2:192.168.123.110:6832/748767188,v1:192.168.123.110:6833/748767188] up:boot 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: fsmap foofs:1 {0=foofs.vm08.jgklas=up:active} 1 up:standby 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "mds metadata", "who": "foofs.vm10.uuhfwt"}]: dispatch 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: fsmap foofs:1 {0=foofs.vm08.jgklas=up:active} 1 up:standby 2026-03-09T13:56:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:36 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:36.923 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- ceph orch ls -f json 2026-03-09T13:56:37.423 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:37 vm08 ceph-mon[47734]: from='client.24343 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:37.423 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:37.423 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:37.423 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:37.423 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:37.423 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:37.424 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:37 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:37.425 INFO:teuthology.orchestra.run.vm08.stdout: 2026-03-09T13:56:37.425 INFO:teuthology.orchestra.run.vm08.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-09T13:53:55.374593Z", "last_refresh": "2026-03-09T13:56:36.757383Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-09T13:53:53.861815Z", "last_refresh": "2026-03-09T13:56:36.172757Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-09T13:53:54.676006Z", "last_refresh": "2026-03-09T13:56:36.757439Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-09T13:56:33.247282Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-09T13:56:33.240629Z", "last_refresh": "2026-03-09T13:56:36.173125Z", "running": 2, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-09T13:53:53.550627Z", "last_refresh": "2026-03-09T13:56:36.172910Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm08:192.168.123.108=vm08", "vm10:192.168.123.110=vm10"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-09T13:54:10.434622Z", "last_refresh": "2026-03-09T13:56:36.172947Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-09T13:53:55.049639Z", "last_refresh": "2026-03-09T13:56:36.172978Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-09T13:56:36.173008Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-09T13:53:54.164175Z", "last_refresh": "2026-03-09T13:56:36.757492Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-09T13:56:37.502 INFO:tasks.cephadm:mds.foofs has 2/2 2026-03-09T13:56:37.502 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T13:56:37.505 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm08.local 2026-03-09T13:56:37.505 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'ceph nfs cluster create foo --placement=2 || ceph nfs cluster create cephfs foo --placement=2' 2026-03-09T13:56:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:37 vm10 ceph-mon[51861]: from='client.24343 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:37 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:38.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:38 vm08 ceph-mon[47734]: pgmap v95: 65 pgs: 20 creating+peering, 9 unknown, 36 active+clean; 449 KiB data, 49 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s wr, 1 op/s 2026-03-09T13:56:38.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:38 vm08 ceph-mon[47734]: from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:38.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:38 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool create", "pool": ".nfs"}]: dispatch 2026-03-09T13:56:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:38 vm10 ceph-mon[51861]: pgmap v95: 65 pgs: 20 creating+peering, 9 unknown, 36 active+clean; 449 KiB data, 49 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s wr, 1 op/s 2026-03-09T13:56:38.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:38 vm10 ceph-mon[51861]: from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T13:56:38.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:38 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool create", "pool": ".nfs"}]: dispatch 2026-03-09T13:56:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:39 vm08 ceph-mon[47734]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "foo", "placement": "2", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:39 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool create", "pool": ".nfs"}]': finished 2026-03-09T13:56:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:39 vm08 ceph-mon[47734]: osdmap e51: 8 total, 8 up, 8 in 2026-03-09T13:56:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:39 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-09T13:56:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:39 vm10 ceph-mon[51861]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "foo", "placement": "2", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:39 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool create", "pool": ".nfs"}]': finished 2026-03-09T13:56:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:39 vm10 ceph-mon[51861]: osdmap e51: 8 total, 8 up, 8 in 2026-03-09T13:56:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:39 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-09T13:56:39.850 INFO:teuthology.orchestra.run.vm08.stdout:NFS Cluster Created Successfully 2026-03-09T13:56:39.921 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'ceph nfs export create cephfs --fsname foofs --clusterid foo --binding /fake || ceph nfs export create cephfs --fsname foofs --cluster-id foo --pseudo-path /fake' 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: pgmap v97: 97 pgs: 32 unknown, 65 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s wr, 5 op/s 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: osdmap e52: 8 total, 8 up, 8 in 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm08.gfvbom", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm08.gfvbom", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm08.gfvbom-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm08.gfvbom-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T13:56:40.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:40 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:40.554 INFO:teuthology.orchestra.run.vm08.stderr:Invalid command: Unexpected argument '--clusterid' 2026-03-09T13:56:40.554 INFO:teuthology.orchestra.run.vm08.stderr:nfs export create cephfs [] [--readonly] [--client_addr ...] [--squash ] : Create a CephFS export 2026-03-09T13:56:40.554 INFO:teuthology.orchestra.run.vm08.stderr:Error EINVAL: invalid command 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: pgmap v97: 97 pgs: 32 unknown, 65 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s wr, 5 op/s 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: osdmap e52: 8 total, 8 up, 8 in 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm08.gfvbom", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-09T13:56:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm08.gfvbom", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm08.gfvbom-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm08.gfvbom-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T13:56:40.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:40 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:40.854 INFO:teuthology.orchestra.run.vm08.stdout:{ 2026-03-09T13:56:40.854 INFO:teuthology.orchestra.run.vm08.stdout: "bind": "/fake", 2026-03-09T13:56:40.854 INFO:teuthology.orchestra.run.vm08.stdout: "fs": "foofs", 2026-03-09T13:56:40.854 INFO:teuthology.orchestra.run.vm08.stdout: "path": "/", 2026-03-09T13:56:40.854 INFO:teuthology.orchestra.run.vm08.stdout: "cluster": "foo", 2026-03-09T13:56:40.854 INFO:teuthology.orchestra.run.vm08.stdout: "mode": "RW" 2026-03-09T13:56:40.854 INFO:teuthology.orchestra.run.vm08.stdout:} 2026-03-09T13:56:41.012 DEBUG:teuthology.orchestra.run.vm08:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4a3e67a6-1bbf-11f1-aca1-21c5f680cc64 -- bash -c 'while ! ceph orch ls | grep nfs | grep 2/2 ; do sleep 1 ; done' 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: Saving service nfs.foo spec with placement count:2 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: Creating key for client.nfs.foo.0.0.vm08.gfvbom 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: Ensuring nfs.foo.0 is in the ganesha grace table 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: Rados config object exists: conf-nfs.foo 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: Creating key for client.nfs.foo.0.0.vm08.gfvbom-rgw 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: Deploying daemon nfs.foo.0.0.vm08.gfvbom on vm08 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: osdmap e53: 8 total, 8 up, 8 in 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm10.quftgv", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm10.quftgv", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-09T13:56:41.270 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:41 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:41.551 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: Saving service nfs.foo spec with placement count:2 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: Creating key for client.nfs.foo.0.0.vm08.gfvbom 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: Ensuring nfs.foo.0 is in the ganesha grace table 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: Rados config object exists: conf-nfs.foo 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: Creating key for client.nfs.foo.0.0.vm08.gfvbom-rgw 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: Deploying daemon nfs.foo.0.0.vm08.gfvbom on vm08 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: osdmap e53: 8 total, 8 up, 8 in 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm10.quftgv", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm10.quftgv", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-09T13:56:41.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:41 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: from='client.24359 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "foofs", "cluster_id": "foo", "pseudo_path": "/fake", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: pgmap v100: 97 pgs: 3 creating+peering, 29 unknown, 65 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s wr, 7 op/s 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: Creating key for client.nfs.foo.1.0.vm10.quftgv 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: Ensuring nfs.foo.1 is in the ganesha grace table 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: Rados config object exists: conf-nfs.foo 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: Creating key for client.nfs.foo.1.0.vm10.quftgv-rgw 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm10.quftgv-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm10.quftgv-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: Deploying daemon nfs.foo.1.0.vm10.quftgv on vm10 2026-03-09T13:56:42.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:42 vm10 ceph-mon[51861]: mgrmap e20: vm08.siqces(active, since 118s), standbys: vm10.euefis 2026-03-09T13:56:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: from='client.24359 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "foofs", "cluster_id": "foo", "pseudo_path": "/fake", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: pgmap v100: 97 pgs: 3 creating+peering, 29 unknown, 65 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s wr, 7 op/s 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: Creating key for client.nfs.foo.1.0.vm10.quftgv 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: Ensuring nfs.foo.1 is in the ganesha grace table 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: Rados config object exists: conf-nfs.foo 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: Creating key for client.nfs.foo.1.0.vm10.quftgv-rgw 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm10.quftgv-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm10.quftgv-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: Deploying daemon nfs.foo.1.0.vm10.quftgv on vm10 2026-03-09T13:56:42.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:42 vm08 ceph-mon[47734]: mgrmap e20: vm08.siqces(active, since 118s), standbys: vm10.euefis 2026-03-09T13:56:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:43 vm08 ceph-mon[47734]: from='client.14604 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:56:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:56:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:43 vm10 ceph-mon[51861]: from='client.14604 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:56:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:56:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.051 INFO:teuthology.orchestra.run.vm08.stdout:nfs.foo ?:2049 2/2 0s ago 4s count:2 2026-03-09T13:56:44.114 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-09T13:56:44.117 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm08.local 2026-03-09T13:56:44.117 DEBUG:teuthology.orchestra.run.vm08:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'mkdir /mnt/foo' 2026-03-09T13:56:44.140 INFO:teuthology.orchestra.run.vm08.stderr:+ mkdir /mnt/foo 2026-03-09T13:56:44.142 DEBUG:teuthology.orchestra.run.vm08:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'while ! mount -t nfs $(hostname):/fake /mnt/foo -o sync ; do sleep 5 ; done' 2026-03-09T13:56:44.205 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:56:44.206 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:56:44.417 INFO:teuthology.orchestra.run.vm08.stderr:Created symlink /run/systemd/system/remote-fs.target.wants/rpc-statd.service → /usr/lib/systemd/system/rpc-statd.service. 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='client.14612 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: pgmap v101: 97 pgs: 3 creating+peering, 10 unknown, 84 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 853 B/s wr, 6 op/s 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.662 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='client.14612 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: pgmap v101: 97 pgs: 3 creating+peering, 10 unknown, 84 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 853 B/s wr, 6 op/s 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:44.713 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:45 vm08 ceph-mon[47734]: from='client.24393 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:45 vm10 ceph-mon[51861]: from='client.24393 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T13:56:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:46 vm08 ceph-mon[47734]: pgmap v102: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 3.1 KiB/s rd, 2.0 KiB/s wr, 8 op/s 2026-03-09T13:56:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:46 vm10 ceph-mon[51861]: pgmap v102: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 3.1 KiB/s rd, 2.0 KiB/s wr, 8 op/s 2026-03-09T13:56:47.895 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:56:47.896 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:56:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:48 vm08 ceph-mon[47734]: pgmap v103: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 1.5 KiB/s wr, 6 op/s 2026-03-09T13:56:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:48 vm10 ceph-mon[51861]: pgmap v103: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 1.5 KiB/s wr, 6 op/s 2026-03-09T13:56:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:50 vm08 ceph-mon[47734]: pgmap v104: 97 pgs: 97 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.5 KiB/s rd, 2.6 KiB/s wr, 7 op/s 2026-03-09T13:56:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:50.813 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:50 vm10 ceph-mon[51861]: pgmap v104: 97 pgs: 97 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.5 KiB/s rd, 2.6 KiB/s wr, 7 op/s 2026-03-09T13:56:50.813 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:56:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:52 vm08 ceph-mon[47734]: pgmap v105: 97 pgs: 97 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.2 KiB/s rd, 2.4 KiB/s wr, 6 op/s 2026-03-09T13:56:52.897 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:56:52.897 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:56:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:52 vm10 ceph-mon[51861]: pgmap v105: 97 pgs: 97 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.2 KiB/s rd, 2.4 KiB/s wr, 6 op/s 2026-03-09T13:56:52.923 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:56:52.923 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:56:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:54 vm08 ceph-mon[47734]: pgmap v106: 97 pgs: 97 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 2.0 KiB/s wr, 5 op/s 2026-03-09T13:56:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:54 vm10 ceph-mon[51861]: pgmap v106: 97 pgs: 97 active+clean; 452 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 2.0 KiB/s wr, 5 op/s 2026-03-09T13:56:56.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:55 vm10 ceph-mon[51861]: pgmap v107: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 2.4 KiB/s wr, 5 op/s 2026-03-09T13:56:56.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:55 vm08 ceph-mon[47734]: pgmap v107: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 2.4 KiB/s wr, 5 op/s 2026-03-09T13:56:57.925 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:56:57.925 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:56:57.951 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:56:57.951 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:56:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:57 vm08 ceph-mon[47734]: pgmap v108: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1.4 KiB/s wr, 2 op/s 2026-03-09T13:56:58.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:57 vm10 ceph-mon[51861]: pgmap v108: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1.4 KiB/s wr, 2 op/s 2026-03-09T13:57:00.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:56:59 vm08 ceph-mon[47734]: pgmap v109: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1.5 KiB/s wr, 3 op/s 2026-03-09T13:57:00.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:56:59 vm10 ceph-mon[51861]: pgmap v109: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1.5 KiB/s wr, 3 op/s 2026-03-09T13:57:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:01 vm08 ceph-mon[47734]: pgmap v110: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 511 B/s wr, 0 op/s 2026-03-09T13:57:02.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:01 vm10 ceph-mon[51861]: pgmap v110: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 511 B/s wr, 0 op/s 2026-03-09T13:57:02.953 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:02.953 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:02.991 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:02.992 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:04.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:04 vm08 ceph-mon[47734]: pgmap v111: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 511 B/s wr, 0 op/s 2026-03-09T13:57:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:04 vm10 ceph-mon[51861]: pgmap v111: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 511 B/s wr, 0 op/s 2026-03-09T13:57:06.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:05 vm10 ceph-mon[51861]: pgmap v112: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 597 B/s wr, 1 op/s 2026-03-09T13:57:06.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:05 vm08 ceph-mon[47734]: pgmap v112: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 597 B/s wr, 1 op/s 2026-03-09T13:57:07.993 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:07.994 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:08.020 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:08.021 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:08.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:07 vm08 ceph-mon[47734]: pgmap v113: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:08.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:07 vm10 ceph-mon[51861]: pgmap v113: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:10.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:09 vm08 ceph-mon[47734]: pgmap v114: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:10.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:09 vm10 ceph-mon[51861]: pgmap v114: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:12 vm08 ceph-mon[47734]: pgmap v115: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:12.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:12 vm10 ceph-mon[51861]: pgmap v115: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:13.022 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:13.022 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:13.048 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:13.049 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:14.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:14 vm08 ceph-mon[47734]: pgmap v116: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:14.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:14 vm10 ceph-mon[51861]: pgmap v116: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:15 vm10 ceph-mon[51861]: pgmap v117: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:16.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:15 vm08 ceph-mon[47734]: pgmap v117: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:18.050 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:18.051 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:18.079 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:18.079 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:18.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:17 vm08 ceph-mon[47734]: pgmap v118: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:18.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:17 vm10 ceph-mon[51861]: pgmap v118: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:20.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:19 vm08 ceph-mon[47734]: pgmap v119: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:20.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:19 vm10 ceph-mon[51861]: pgmap v119: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:22.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:21 vm08 ceph-mon[47734]: pgmap v120: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:22.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:21 vm10 ceph-mon[51861]: pgmap v120: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:23.081 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:23.081 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:23.108 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:23.109 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:24.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:24 vm08 ceph-mon[47734]: pgmap v121: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:24 vm10 ceph-mon[51861]: pgmap v121: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:26.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:25 vm10 ceph-mon[51861]: pgmap v122: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:26.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:25 vm08 ceph-mon[47734]: pgmap v122: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:28.110 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:28.111 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:28.138 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:28.139 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:28.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:27 vm08 ceph-mon[47734]: pgmap v123: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:28.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:27 vm10 ceph-mon[51861]: pgmap v123: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:29 vm08 ceph-mon[47734]: pgmap v124: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:29 vm10 ceph-mon[51861]: pgmap v124: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:32.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:32 vm10 ceph-mon[51861]: pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:32 vm08 ceph-mon[47734]: pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:33.140 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:33.141 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:33.166 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:33.167 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:34 vm08 ceph-mon[47734]: pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:34 vm10 ceph-mon[51861]: pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:36.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:35 vm10 ceph-mon[51861]: pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:36.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:35 vm08 ceph-mon[47734]: pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:38.168 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:38.168 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:38.195 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:38.195 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:38.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:37 vm08 ceph-mon[47734]: pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:38.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:37 vm10 ceph-mon[51861]: pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:40.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:40 vm10 ceph-mon[51861]: pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:40 vm08 ceph-mon[47734]: pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:42 vm08 ceph-mon[47734]: pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:42 vm10 ceph-mon[51861]: pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:43.197 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:43.197 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:43.229 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:43.229 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:43.521 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.f", "id": [7, 2]}]: dispatch 2026-03-09T13:57:43.521 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-09T13:57:43.522 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.6", "id": [1, 5]}]: dispatch 2026-03-09T13:57:43.522 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 2]}]: dispatch 2026-03-09T13:57:43.522 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:57:43.522 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:57:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.f", "id": [7, 2]}]: dispatch 2026-03-09T13:57:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-09T13:57:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.6", "id": [1, 5]}]: dispatch 2026-03-09T13:57:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 2]}]: dispatch 2026-03-09T13:57:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:57:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:57:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: pgmap v131: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.f", "id": [7, 2]}]': finished 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]': finished 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.6", "id": [1, 5]}]': finished 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 2]}]': finished 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: osdmap e54: 8 total, 8 up, 8 in 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:57:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: pgmap v131: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.f", "id": [7, 2]}]': finished 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]': finished 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.6", "id": [1, 5]}]': finished 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.18", "id": [1, 2]}]': finished 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: osdmap e54: 8 total, 8 up, 8 in 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:57:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:57:45.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:45 vm08 ceph-mon[47734]: osdmap e55: 8 total, 8 up, 8 in 2026-03-09T13:57:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:45 vm10 ceph-mon[51861]: osdmap e55: 8 total, 8 up, 8 in 2026-03-09T13:57:46.480 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:46 vm08 ceph-mon[47734]: pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:46 vm10 ceph-mon[51861]: pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:57:48.230 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:48.231 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:48.258 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:48.258 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:48 vm08 ceph-mon[47734]: pgmap v135: 97 pgs: 1 active+recovering, 96 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 127 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:48 vm10 ceph-mon[51861]: pgmap v135: 97 pgs: 1 active+recovering, 96 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 127 B/s rd, 127 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:50 vm08 ceph-mon[47734]: pgmap v136: 97 pgs: 1 active+recovering, 96 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:50 vm10 ceph-mon[51861]: pgmap v136: 97 pgs: 1 active+recovering, 96 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:52 vm08 ceph-mon[47734]: pgmap v137: 97 pgs: 1 active+recovering, 96 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:52 vm10 ceph-mon[51861]: pgmap v137: 97 pgs: 1 active+recovering, 96 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:53.260 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:53.260 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:53.284 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:53.285 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:54 vm08 ceph-mon[47734]: pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 210 B/s rd, 210 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:54 vm10 ceph-mon[51861]: pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 210 B/s rd, 210 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:56.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:55 vm10 ceph-mon[51861]: pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 190 B/s rd, 190 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:56.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:55 vm08 ceph-mon[47734]: pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 190 B/s rd, 190 B/s wr, 0 op/s; 2 B/s, 0 objects/s recovering 2026-03-09T13:57:58.286 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:57:58.287 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:57:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:57 vm08 ceph-mon[47734]: pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s; 1 B/s, 0 objects/s recovering 2026-03-09T13:57:58.311 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:57:58.312 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:57:58.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:57 vm10 ceph-mon[51861]: pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s; 1 B/s, 0 objects/s recovering 2026-03-09T13:58:00.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:57:59 vm08 ceph-mon[47734]: pgmap v141: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:00.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:57:59 vm10 ceph-mon[51861]: pgmap v141: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:02 vm08 ceph-mon[47734]: pgmap v142: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:02.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:02 vm10 ceph-mon[51861]: pgmap v142: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:03.313 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:03.313 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:03.339 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:03.339 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:04.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:04 vm08 ceph-mon[47734]: pgmap v143: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:04 vm10 ceph-mon[51861]: pgmap v143: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:06.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:06 vm08 ceph-mon[47734]: pgmap v144: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:06.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:06 vm10 ceph-mon[51861]: pgmap v144: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:08.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:08 vm08 ceph-mon[47734]: pgmap v145: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:08.340 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:08.341 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:08.366 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:08.367 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:08 vm10 ceph-mon[51861]: pgmap v145: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:10.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:10 vm08 ceph-mon[47734]: pgmap v146: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:10.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:10 vm10 ceph-mon[51861]: pgmap v146: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:12 vm08 ceph-mon[47734]: pgmap v147: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:12.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:12 vm10 ceph-mon[51861]: pgmap v147: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:13.368 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:13.368 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:13.393 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:13.394 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:14.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:14 vm10 ceph-mon[51861]: pgmap v148: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:14 vm08 ceph-mon[47734]: pgmap v148: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:16.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:16 vm10 ceph-mon[51861]: pgmap v149: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:16.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:16 vm08 ceph-mon[47734]: pgmap v149: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:18.395 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:18.396 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:18.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:18 vm10 ceph-mon[51861]: pgmap v150: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:18.422 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:18.422 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:18 vm08 ceph-mon[47734]: pgmap v150: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:20 vm08 ceph-mon[47734]: pgmap v151: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:20 vm10 ceph-mon[51861]: pgmap v151: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:22 vm10 ceph-mon[51861]: pgmap v152: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:22.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:22 vm08 ceph-mon[47734]: pgmap v152: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:23.424 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:23.424 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:23.449 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:23.450 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:24 vm10 ceph-mon[51861]: pgmap v153: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:24 vm08 ceph-mon[47734]: pgmap v153: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:26 vm10 ceph-mon[51861]: pgmap v154: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:26 vm08 ceph-mon[47734]: pgmap v154: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:28.451 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:28.452 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:28.491 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:28.492 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:28 vm10 ceph-mon[51861]: pgmap v155: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:28.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:28 vm08 ceph-mon[47734]: pgmap v155: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:30 vm08 ceph-mon[47734]: pgmap v156: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:30 vm10 ceph-mon[51861]: pgmap v156: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:32 vm08 ceph-mon[47734]: pgmap v157: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:32 vm10 ceph-mon[51861]: pgmap v157: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:33.493 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:33.494 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:33.518 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:33.519 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:34 vm08 ceph-mon[47734]: pgmap v158: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:34 vm10 ceph-mon[51861]: pgmap v158: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:36 vm08 ceph-mon[47734]: pgmap v159: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:36 vm10 ceph-mon[51861]: pgmap v159: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:38.521 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:38.521 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:38.547 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:38.548 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:38 vm08 ceph-mon[47734]: pgmap v160: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:38 vm10 ceph-mon[51861]: pgmap v160: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:40 vm08 ceph-mon[47734]: pgmap v161: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:40 vm10 ceph-mon[51861]: pgmap v161: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:42 vm10 ceph-mon[51861]: pgmap v162: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:42 vm08 ceph-mon[47734]: pgmap v162: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:43.549 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:43.550 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:43.575 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:43.576 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:58:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:58:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:58:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:58:44.571 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:44 vm10 ceph-mon[51861]: pgmap v163: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:44.571 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:58:44.571 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:58:44.571 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:58:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:58:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:44 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:58:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:44 vm08 ceph-mon[47734]: pgmap v163: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:58:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:58:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:58:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:58:45.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:44 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:58:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:46 vm08 ceph-mon[47734]: pgmap v164: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:46 vm10 ceph-mon[51861]: pgmap v164: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:48.577 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:48.578 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:48.607 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:48.608 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:48 vm10 ceph-mon[51861]: pgmap v165: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:48 vm08 ceph-mon[47734]: pgmap v165: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:50 vm10 ceph-mon[51861]: pgmap v166: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:50 vm08 ceph-mon[47734]: pgmap v166: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:52 vm10 ceph-mon[51861]: pgmap v167: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:52 vm08 ceph-mon[47734]: pgmap v167: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:53.609 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:53.609 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:53.635 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:53.636 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:58:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:54 vm10 ceph-mon[51861]: pgmap v168: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:54 vm08 ceph-mon[47734]: pgmap v168: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:56.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:56 vm08 ceph-mon[47734]: pgmap v169: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:56 vm10 ceph-mon[51861]: pgmap v169: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:58:57.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:58:57 vm10 ceph-mon[51861]: pgmap v170: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:58.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:58:57 vm08 ceph-mon[47734]: pgmap v170: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:58:58.637 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:58:58.638 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:58:58.664 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:58:58.665 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:00.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:00 vm08 ceph-mon[47734]: pgmap v171: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:00.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:00 vm10 ceph-mon[51861]: pgmap v171: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:02.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:02 vm10 ceph-mon[51861]: pgmap v172: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:02 vm08 ceph-mon[47734]: pgmap v172: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:03.666 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:03.667 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:03.693 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:03.694 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:04 vm10 ceph-mon[51861]: pgmap v173: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:04 vm08 ceph-mon[47734]: pgmap v173: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:06.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:06 vm10 ceph-mon[51861]: pgmap v174: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:06 vm08 ceph-mon[47734]: pgmap v174: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:08.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:08 vm10 ceph-mon[51861]: pgmap v175: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T13:59:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:08 vm08 ceph-mon[47734]: pgmap v175: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T13:59:08.695 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:08.696 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:08.721 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:08.721 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:10.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:10 vm10 ceph-mon[51861]: pgmap v176: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:10 vm08 ceph-mon[47734]: pgmap v176: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:12 vm08 ceph-mon[47734]: pgmap v177: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T13:59:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:12 vm10 ceph-mon[51861]: pgmap v177: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T13:59:13.723 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:13.723 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:13.748 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:13.749 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:14 vm08 ceph-mon[47734]: pgmap v178: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T13:59:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:14 vm10 ceph-mon[51861]: pgmap v178: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T13:59:16.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:16 vm08 ceph-mon[47734]: pgmap v179: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:16 vm10 ceph-mon[51861]: pgmap v179: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:18 vm08 ceph-mon[47734]: pgmap v180: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:18 vm10 ceph-mon[51861]: pgmap v180: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:18.750 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:18.750 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:18.775 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:18.776 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:20 vm08 ceph-mon[47734]: pgmap v181: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:20 vm10 ceph-mon[51861]: pgmap v181: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:22 vm08 ceph-mon[47734]: pgmap v182: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:22 vm10 ceph-mon[51861]: pgmap v182: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:23.777 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:23.778 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:23.803 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:23.803 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:24 vm08 ceph-mon[47734]: pgmap v183: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:24 vm10 ceph-mon[51861]: pgmap v183: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:26.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:26 vm08 ceph-mon[47734]: pgmap v184: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:26 vm10 ceph-mon[51861]: pgmap v184: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:28 vm08 ceph-mon[47734]: pgmap v185: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:28 vm10 ceph-mon[51861]: pgmap v185: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:28.805 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:28.805 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:28.831 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:28.832 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:30 vm08 ceph-mon[47734]: pgmap v186: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:30 vm10 ceph-mon[51861]: pgmap v186: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:32 vm08 ceph-mon[47734]: pgmap v187: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:32 vm10 ceph-mon[51861]: pgmap v187: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:33.833 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:33.834 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:33.862 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:33.862 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:34 vm08 ceph-mon[47734]: pgmap v188: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:34 vm10 ceph-mon[51861]: pgmap v188: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:36.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:36 vm08 ceph-mon[47734]: pgmap v189: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:36 vm10 ceph-mon[51861]: pgmap v189: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:38 vm08 ceph-mon[47734]: pgmap v190: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:38 vm10 ceph-mon[51861]: pgmap v190: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:38.863 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:38.864 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:38.889 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:38.889 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:40 vm08 ceph-mon[47734]: pgmap v191: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:40 vm10 ceph-mon[51861]: pgmap v191: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:42 vm08 ceph-mon[47734]: pgmap v192: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:42 vm10 ceph-mon[51861]: pgmap v192: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:43.531 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:59:43.531 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:59:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T13:59:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T13:59:43.891 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:43.891 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:43.916 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:43.917 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:44 vm08 ceph-mon[47734]: pgmap v193: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:44.588 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:44 vm10 ceph-mon[51861]: pgmap v193: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:59:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:59:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:59:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:59:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:59:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T13:59:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T13:59:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T13:59:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:59:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T13:59:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:46 vm08 ceph-mon[47734]: pgmap v194: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:46 vm10 ceph-mon[51861]: pgmap v194: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:47.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:47 vm10 ceph-mon[51861]: pgmap v195: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:48.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:47 vm08 ceph-mon[47734]: pgmap v195: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:48.918 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:48.919 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:48.945 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:48.945 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:50.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:50 vm08 ceph-mon[47734]: pgmap v196: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:50.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:50 vm10 ceph-mon[51861]: pgmap v196: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:52.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:52 vm10 ceph-mon[51861]: pgmap v197: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:52 vm08 ceph-mon[47734]: pgmap v197: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:53.947 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:53.947 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:53.972 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:53.972 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T13:59:54.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:54 vm10 ceph-mon[51861]: pgmap v198: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:54 vm08 ceph-mon[47734]: pgmap v198: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:56.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:56 vm10 ceph-mon[51861]: pgmap v199: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:56 vm08 ceph-mon[47734]: pgmap v199: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T13:59:58.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 13:59:58 vm10 ceph-mon[51861]: pgmap v200: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 13:59:58 vm08 ceph-mon[47734]: pgmap v200: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T13:59:58.974 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T13:59:58.974 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T13:59:59.001 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T13:59:59.001 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:00 vm08 ceph-mon[47734]: pgmap v201: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:00 vm08 ceph-mon[47734]: overall HEALTH_OK 2026-03-09T14:00:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:00 vm10 ceph-mon[51861]: pgmap v201: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:00 vm10 ceph-mon[51861]: overall HEALTH_OK 2026-03-09T14:00:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:02 vm10 ceph-mon[51861]: pgmap v202: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:02 vm08 ceph-mon[47734]: pgmap v202: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:04.003 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:04.003 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:04.028 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:04.029 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:04 vm10 ceph-mon[51861]: pgmap v203: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:04 vm08 ceph-mon[47734]: pgmap v203: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:06 vm08 ceph-mon[47734]: pgmap v204: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:06 vm10 ceph-mon[51861]: pgmap v204: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:08 vm10 ceph-mon[51861]: pgmap v205: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:09.030 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:09.031 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:09.056 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:08 vm08 ceph-mon[47734]: pgmap v205: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:09.057 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:10 vm10 ceph-mon[51861]: pgmap v206: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:10 vm08 ceph-mon[47734]: pgmap v206: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:12 vm10 ceph-mon[51861]: pgmap v207: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:12 vm08 ceph-mon[47734]: pgmap v207: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:14.058 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:14.059 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:14.087 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:14.088 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:14 vm10 ceph-mon[51861]: pgmap v208: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:14 vm08 ceph-mon[47734]: pgmap v208: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:16 vm08 ceph-mon[47734]: pgmap v209: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:16 vm10 ceph-mon[51861]: pgmap v209: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:18 vm10 ceph-mon[51861]: pgmap v210: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:18 vm08 ceph-mon[47734]: pgmap v210: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:19.089 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:19.090 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:19.116 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:19.117 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:20 vm10 ceph-mon[51861]: pgmap v211: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:20 vm08 ceph-mon[47734]: pgmap v211: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:22 vm10 ceph-mon[51861]: pgmap v212: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:22 vm08 ceph-mon[47734]: pgmap v212: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:23 vm08 ceph-mon[47734]: pgmap v213: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:24.118 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:24.119 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:24.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:23 vm10 ceph-mon[51861]: pgmap v213: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:24.166 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:24.166 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:26 vm10 ceph-mon[51861]: pgmap v214: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:26.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:26 vm08 ceph-mon[47734]: pgmap v214: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:28.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:28 vm10 ceph-mon[51861]: pgmap v215: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:28 vm08 ceph-mon[47734]: pgmap v215: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:29.168 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:29.168 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:29.196 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:29.197 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:30 vm10 ceph-mon[51861]: pgmap v216: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:30 vm08 ceph-mon[47734]: pgmap v216: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:32 vm08 ceph-mon[47734]: pgmap v217: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:32 vm10 ceph-mon[51861]: pgmap v217: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:34.198 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:34.199 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:34.226 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:34.226 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:34 vm08 ceph-mon[47734]: pgmap v218: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:34 vm10 ceph-mon[51861]: pgmap v218: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:36.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:36 vm08 ceph-mon[47734]: pgmap v219: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:36 vm10 ceph-mon[51861]: pgmap v219: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:38 vm08 ceph-mon[47734]: pgmap v220: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:38 vm10 ceph-mon[51861]: pgmap v220: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:39.227 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:39.228 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:39.255 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:39.256 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:40 vm08 ceph-mon[47734]: pgmap v221: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:40 vm10 ceph-mon[51861]: pgmap v221: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:42 vm08 ceph-mon[47734]: pgmap v222: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:42 vm10 ceph-mon[51861]: pgmap v222: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:43.537 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:00:43.538 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:00:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:00:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:00:44.257 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:44.258 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:44.286 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:44.287 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:44 vm08 ceph-mon[47734]: pgmap v223: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:44.638 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:44 vm10 ceph-mon[51861]: pgmap v223: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:00:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:00:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:00:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:00:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:00:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:00:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:00:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:00:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:00:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:00:46.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:46 vm08 ceph-mon[47734]: pgmap v224: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:46 vm10 ceph-mon[51861]: pgmap v224: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:48 vm08 ceph-mon[47734]: pgmap v225: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:48 vm10 ceph-mon[51861]: pgmap v225: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:49.288 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:49.289 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:49.316 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:49.317 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:50 vm08 ceph-mon[47734]: pgmap v226: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:50 vm10 ceph-mon[51861]: pgmap v226: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:52 vm08 ceph-mon[47734]: pgmap v227: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:52 vm10 ceph-mon[51861]: pgmap v227: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:54.318 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:54.319 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:54.347 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:54.348 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:00:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:54 vm08 ceph-mon[47734]: pgmap v228: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:54 vm10 ceph-mon[51861]: pgmap v228: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:56 vm08 ceph-mon[47734]: pgmap v229: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:56 vm10 ceph-mon[51861]: pgmap v229: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:00:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:00:58 vm08 ceph-mon[47734]: pgmap v230: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:00:58 vm10 ceph-mon[51861]: pgmap v230: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:00:59.349 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:00:59.350 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:00:59.376 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:00:59.377 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:00.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:00 vm10 ceph-mon[51861]: pgmap v231: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:00.506 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:00 vm08 ceph-mon[47734]: pgmap v231: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:02 vm08 ceph-mon[47734]: pgmap v232: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:02 vm10 ceph-mon[51861]: pgmap v232: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:04.379 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:04.379 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:04.405 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:04.406 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:04 vm08 ceph-mon[47734]: pgmap v233: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:04 vm10 ceph-mon[51861]: pgmap v233: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:06.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:06 vm08 ceph-mon[47734]: pgmap v234: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:06 vm10 ceph-mon[51861]: pgmap v234: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:08 vm08 ceph-mon[47734]: pgmap v235: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:08 vm10 ceph-mon[51861]: pgmap v235: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:09.407 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:09.408 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:09.434 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:09.434 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:10.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:10 vm10 ceph-mon[51861]: pgmap v236: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:10 vm08 ceph-mon[47734]: pgmap v236: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:12 vm08 ceph-mon[47734]: pgmap v237: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:12 vm10 ceph-mon[51861]: pgmap v237: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:14.436 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:14.436 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:14.461 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:14.462 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:14 vm08 ceph-mon[47734]: pgmap v238: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:14 vm10 ceph-mon[51861]: pgmap v238: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:16.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:16 vm08 ceph-mon[47734]: pgmap v239: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:16 vm10 ceph-mon[51861]: pgmap v239: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:18 vm08 ceph-mon[47734]: pgmap v240: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:18 vm10 ceph-mon[51861]: pgmap v240: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:19.463 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:19.464 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:19.489 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:19.490 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:20 vm08 ceph-mon[47734]: pgmap v241: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:20 vm10 ceph-mon[51861]: pgmap v241: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:22 vm10 ceph-mon[51861]: pgmap v242: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:22.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:22 vm08 ceph-mon[47734]: pgmap v242: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:24.491 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:24.492 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:24.517 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:24.518 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:24 vm10 ceph-mon[51861]: pgmap v243: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:24 vm08 ceph-mon[47734]: pgmap v243: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:26 vm10 ceph-mon[51861]: pgmap v244: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:26 vm08 ceph-mon[47734]: pgmap v244: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:28 vm10 ceph-mon[51861]: pgmap v245: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:28.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:28 vm08 ceph-mon[47734]: pgmap v245: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:29.519 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:29.520 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:29.545 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:29.546 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:30.773 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:30 vm08 ceph-mon[47734]: pgmap v246: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:30 vm10 ceph-mon[51861]: pgmap v246: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:32 vm08 ceph-mon[47734]: pgmap v247: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:32 vm10 ceph-mon[51861]: pgmap v247: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:34.547 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:34.548 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:34.574 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:34.575 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:34 vm08 ceph-mon[47734]: pgmap v248: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:34 vm10 ceph-mon[51861]: pgmap v248: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:36 vm08 ceph-mon[47734]: pgmap v249: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:36 vm10 ceph-mon[51861]: pgmap v249: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:38 vm08 ceph-mon[47734]: pgmap v250: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:38 vm10 ceph-mon[51861]: pgmap v250: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:39.576 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:39.577 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:39.654 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:39.655 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:40 vm08 ceph-mon[47734]: pgmap v251: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:40 vm10 ceph-mon[51861]: pgmap v251: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:42.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:42 vm08 ceph-mon[47734]: pgmap v252: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:42 vm10 ceph-mon[51861]: pgmap v252: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:01:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:01:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:01:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:01:44.656 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:44.657 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:44.729 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:44.729 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:44 vm08 ceph-mon[47734]: pgmap v253: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:44.889 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:44 vm10 ceph-mon[51861]: pgmap v253: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:01:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:01:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:01:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:01:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:01:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:01:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:01:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:01:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:01:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:01:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:46 vm08 ceph-mon[47734]: pgmap v254: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:46 vm10 ceph-mon[51861]: pgmap v254: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:48 vm08 ceph-mon[47734]: pgmap v255: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:48 vm10 ceph-mon[51861]: pgmap v255: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:49.730 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:49.731 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:49.756 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:49.757 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:50 vm08 ceph-mon[47734]: pgmap v256: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:50 vm10 ceph-mon[51861]: pgmap v256: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:52.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:52 vm08 ceph-mon[47734]: pgmap v257: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:52 vm10 ceph-mon[51861]: pgmap v257: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:54.758 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:54.759 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:54.785 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:54.786 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:01:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:54 vm08 ceph-mon[47734]: pgmap v258: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:54 vm10 ceph-mon[51861]: pgmap v258: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:56 vm08 ceph-mon[47734]: pgmap v259: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:56 vm10 ceph-mon[51861]: pgmap v259: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:01:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:01:58 vm08 ceph-mon[47734]: pgmap v260: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:01:58 vm10 ceph-mon[51861]: pgmap v260: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:01:59.787 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:01:59.787 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:01:59.813 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:01:59.814 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:00 vm10 ceph-mon[51861]: pgmap v261: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:00 vm08 ceph-mon[47734]: pgmap v261: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:02 vm10 ceph-mon[51861]: pgmap v262: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:02 vm08 ceph-mon[47734]: pgmap v262: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:04.815 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:04.816 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:04.841 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:04.842 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:04 vm10 ceph-mon[51861]: pgmap v263: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:04 vm08 ceph-mon[47734]: pgmap v263: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:06 vm08 ceph-mon[47734]: pgmap v264: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:06 vm10 ceph-mon[51861]: pgmap v264: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:07.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:07 vm10 ceph-mon[51861]: pgmap v265: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:08.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:07 vm08 ceph-mon[47734]: pgmap v265: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:09.843 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:09.844 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:09.906 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:09.907 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:10.125 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:10 vm10 ceph-mon[51861]: pgmap v266: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:10.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:10 vm08 ceph-mon[47734]: pgmap v266: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:12 vm08 ceph-mon[47734]: pgmap v267: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:12 vm10 ceph-mon[51861]: pgmap v267: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:14 vm08 ceph-mon[47734]: pgmap v268: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:14 vm10 ceph-mon[51861]: pgmap v268: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:14.908 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:14.908 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:14.934 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:14.934 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:16.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:16 vm08 ceph-mon[47734]: pgmap v269: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:16 vm10 ceph-mon[51861]: pgmap v269: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:18 vm08 ceph-mon[47734]: pgmap v270: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:18 vm10 ceph-mon[51861]: pgmap v270: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:19.935 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:19.936 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:19.961 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:19.962 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:20 vm08 ceph-mon[47734]: pgmap v271: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:20 vm10 ceph-mon[51861]: pgmap v271: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:22 vm08 ceph-mon[47734]: pgmap v272: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:22 vm10 ceph-mon[51861]: pgmap v272: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:24 vm08 ceph-mon[47734]: pgmap v273: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:24 vm10 ceph-mon[51861]: pgmap v273: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:24.963 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:24.964 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:24.991 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:24.991 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:26.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:26 vm08 ceph-mon[47734]: pgmap v274: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:26 vm10 ceph-mon[51861]: pgmap v274: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:28 vm08 ceph-mon[47734]: pgmap v275: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:28 vm10 ceph-mon[51861]: pgmap v275: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:29.993 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:29.993 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:30.031 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:30.031 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:30 vm08 ceph-mon[47734]: pgmap v276: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:30 vm10 ceph-mon[51861]: pgmap v276: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:32 vm08 ceph-mon[47734]: pgmap v277: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:32 vm10 ceph-mon[51861]: pgmap v277: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:34 vm08 ceph-mon[47734]: pgmap v278: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:34 vm10 ceph-mon[51861]: pgmap v278: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:35.032 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:35.033 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:35.058 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:35.058 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:36.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:36 vm08 ceph-mon[47734]: pgmap v279: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:36 vm10 ceph-mon[51861]: pgmap v279: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:38 vm10 ceph-mon[51861]: pgmap v280: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:38.665 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:38 vm08 ceph-mon[47734]: pgmap v280: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:40.060 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:40.060 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:40.086 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:40.086 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:40 vm10 ceph-mon[51861]: pgmap v281: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:40 vm08 ceph-mon[47734]: pgmap v281: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:42 vm10 ceph-mon[51861]: pgmap v282: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:42.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:42 vm08 ceph-mon[47734]: pgmap v282: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:02:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:02:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:02:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:02:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:44 vm10 ceph-mon[51861]: pgmap v283: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:44 vm08 ceph-mon[47734]: pgmap v283: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:45.087 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:45.088 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:45.112 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:45.113 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:02:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:02:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:45 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:02:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:02:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:02:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:45 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:02:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:46 vm08 ceph-mon[47734]: pgmap v284: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:02:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:02:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:46 vm10 ceph-mon[51861]: pgmap v284: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:02:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:02:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:48 vm08 ceph-mon[47734]: pgmap v285: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:48 vm10 ceph-mon[51861]: pgmap v285: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:50.114 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:50.114 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:50.139 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:50.140 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:50 vm08 ceph-mon[47734]: pgmap v286: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:50 vm10 ceph-mon[51861]: pgmap v286: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:52 vm08 ceph-mon[47734]: pgmap v287: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:52 vm10 ceph-mon[51861]: pgmap v287: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:54 vm08 ceph-mon[47734]: pgmap v288: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:54 vm10 ceph-mon[51861]: pgmap v288: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:55.141 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:02:55.142 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:02:55.167 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:02:55.167 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:02:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:56 vm08 ceph-mon[47734]: pgmap v289: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:56 vm10 ceph-mon[51861]: pgmap v289: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:02:58.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:02:58 vm08 ceph-mon[47734]: pgmap v290: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:02:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:02:58 vm10 ceph-mon[51861]: pgmap v290: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:00.169 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:00.169 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:00.194 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:00.195 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:00 vm10 ceph-mon[51861]: pgmap v291: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:00 vm08 ceph-mon[47734]: pgmap v291: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:02 vm10 ceph-mon[51861]: pgmap v292: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:02 vm08 ceph-mon[47734]: pgmap v292: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:04 vm10 ceph-mon[51861]: pgmap v293: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:04 vm08 ceph-mon[47734]: pgmap v293: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:05.196 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:05.197 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:05.223 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:05.223 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:06 vm08 ceph-mon[47734]: pgmap v294: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:06 vm10 ceph-mon[51861]: pgmap v294: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:08 vm10 ceph-mon[51861]: pgmap v295: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:08 vm08 ceph-mon[47734]: pgmap v295: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:10.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:09 vm08 ceph-mon[47734]: pgmap v296: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:10.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:09 vm10 ceph-mon[51861]: pgmap v296: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:10.224 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:10.225 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:10.251 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:10.252 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:12 vm08 ceph-mon[47734]: pgmap v297: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:12 vm10 ceph-mon[51861]: pgmap v297: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:14 vm08 ceph-mon[47734]: pgmap v298: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:14 vm10 ceph-mon[51861]: pgmap v298: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:15.253 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:15.253 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:15.286 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:15.287 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:16.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:16 vm08 ceph-mon[47734]: pgmap v299: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:16 vm10 ceph-mon[51861]: pgmap v299: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:18 vm08 ceph-mon[47734]: pgmap v300: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:18 vm10 ceph-mon[51861]: pgmap v300: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:20.288 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:20.288 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:20.314 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:20.315 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:20 vm08 ceph-mon[47734]: pgmap v301: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:20 vm10 ceph-mon[51861]: pgmap v301: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:22 vm10 ceph-mon[51861]: pgmap v302: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:22.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:22 vm08 ceph-mon[47734]: pgmap v302: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:24 vm10 ceph-mon[51861]: pgmap v303: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:24 vm08 ceph-mon[47734]: pgmap v303: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:25.316 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:25.317 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:25.349 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:25.349 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:26 vm10 ceph-mon[51861]: pgmap v304: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:26 vm08 ceph-mon[47734]: pgmap v304: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:28 vm10 ceph-mon[51861]: pgmap v305: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:28.669 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:28 vm08 ceph-mon[47734]: pgmap v305: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:30.351 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:30.351 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:30.377 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:30.377 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:30.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:30 vm08 ceph-mon[47734]: pgmap v306: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:30 vm10 ceph-mon[51861]: pgmap v306: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:32 vm08 ceph-mon[47734]: pgmap v307: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:32 vm10 ceph-mon[51861]: pgmap v307: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:34 vm08 ceph-mon[47734]: pgmap v308: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:34 vm10 ceph-mon[51861]: pgmap v308: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:35.379 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:35.379 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:35.404 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:35.404 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:36 vm08 ceph-mon[47734]: pgmap v309: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:36 vm10 ceph-mon[51861]: pgmap v309: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:38 vm08 ceph-mon[47734]: pgmap v310: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:38 vm10 ceph-mon[51861]: pgmap v310: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:40.406 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:40.406 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:40.432 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:40.432 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:40.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:40 vm08 ceph-mon[47734]: pgmap v311: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:40 vm10 ceph-mon[51861]: pgmap v311: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:42.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:42 vm08 ceph-mon[47734]: pgmap v312: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:42 vm10 ceph-mon[51861]: pgmap v312: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:03:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:03:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:03:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:03:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:44 vm08 ceph-mon[47734]: pgmap v313: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:44 vm10 ceph-mon[51861]: pgmap v313: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:45.434 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:45.435 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:45.475 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:45.475 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:46 vm08 ceph-mon[47734]: pgmap v314: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:03:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:03:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:03:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:03:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:03:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:46 vm10 ceph-mon[51861]: pgmap v314: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:03:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:03:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:03:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:03:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:03:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:48 vm08 ceph-mon[47734]: pgmap v315: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:48 vm10 ceph-mon[51861]: pgmap v315: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:50.476 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:50.481 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:50.557 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:50.557 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:50 vm08 ceph-mon[47734]: pgmap v316: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:50 vm10 ceph-mon[51861]: pgmap v316: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:52 vm08 ceph-mon[47734]: pgmap v317: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:52 vm10 ceph-mon[51861]: pgmap v317: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:54 vm08 ceph-mon[47734]: pgmap v318: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:54 vm10 ceph-mon[51861]: pgmap v318: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:55.558 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:03:55.559 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:03:55.613 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:03:55.613 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:03:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:56 vm08 ceph-mon[47734]: pgmap v319: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:56 vm10 ceph-mon[51861]: pgmap v319: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:03:58.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:03:58 vm08 ceph-mon[47734]: pgmap v320: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:03:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:03:58 vm10 ceph-mon[51861]: pgmap v320: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:00.615 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:00.615 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:00.642 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:00.643 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:00 vm08 ceph-mon[47734]: pgmap v321: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:00 vm10 ceph-mon[51861]: pgmap v321: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:02 vm10 ceph-mon[51861]: pgmap v322: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:02 vm08 ceph-mon[47734]: pgmap v322: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:04 vm10 ceph-mon[51861]: pgmap v323: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:04 vm08 ceph-mon[47734]: pgmap v323: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:05.644 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:05.644 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:05.675 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:05.676 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:06 vm08 ceph-mon[47734]: pgmap v324: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:06 vm10 ceph-mon[51861]: pgmap v324: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:08 vm10 ceph-mon[51861]: pgmap v325: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:08 vm08 ceph-mon[47734]: pgmap v325: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:10.677 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:10.677 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:10.705 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:10.705 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:10 vm10 ceph-mon[51861]: pgmap v326: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:10 vm08 ceph-mon[47734]: pgmap v326: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:12.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:11 vm08 ceph-mon[47734]: pgmap v327: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:11 vm10 ceph-mon[51861]: pgmap v327: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:14.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:14 vm10 ceph-mon[51861]: pgmap v328: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:14 vm08 ceph-mon[47734]: pgmap v328: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:15.707 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:15.707 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:15.735 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:15.735 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:16 vm10 ceph-mon[51861]: pgmap v329: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:16 vm08 ceph-mon[47734]: pgmap v329: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:18 vm10 ceph-mon[51861]: pgmap v330: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:18.674 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:18 vm08 ceph-mon[47734]: pgmap v330: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:20 vm10 ceph-mon[51861]: pgmap v331: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:20.736 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:20.737 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:20.762 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:20.763 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:20 vm08 ceph-mon[47734]: pgmap v331: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:22 vm10 ceph-mon[51861]: pgmap v332: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:22.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:22 vm08 ceph-mon[47734]: pgmap v332: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:24 vm10 ceph-mon[51861]: pgmap v333: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:24 vm08 ceph-mon[47734]: pgmap v333: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:25.764 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:25.765 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:25.793 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:25.793 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:26 vm10 ceph-mon[51861]: pgmap v334: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:26 vm08 ceph-mon[47734]: pgmap v334: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:28 vm10 ceph-mon[51861]: pgmap v335: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:28.674 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:28 vm08 ceph-mon[47734]: pgmap v335: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:30 vm10 ceph-mon[51861]: pgmap v336: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:30.794 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:30.795 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:30 vm08 ceph-mon[47734]: pgmap v336: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:30.823 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:30.823 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:32 vm10 ceph-mon[51861]: pgmap v337: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:32 vm08 ceph-mon[47734]: pgmap v337: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:34 vm10 ceph-mon[51861]: pgmap v338: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:34.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:34 vm08 ceph-mon[47734]: pgmap v338: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:35.825 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:35.826 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:35.852 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:35.852 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:36 vm10 ceph-mon[51861]: pgmap v339: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:36 vm08 ceph-mon[47734]: pgmap v339: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:38 vm10 ceph-mon[51861]: pgmap v340: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:38.675 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:38 vm08 ceph-mon[47734]: pgmap v340: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:40 vm08 ceph-mon[47734]: pgmap v341: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:40.853 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:40.854 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:40.879 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:40.880 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:40 vm10 ceph-mon[51861]: pgmap v341: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:42.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:42 vm08 ceph-mon[47734]: pgmap v342: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:42 vm10 ceph-mon[51861]: pgmap v342: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:04:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:04:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:04:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:04:44.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:44 vm08 ceph-mon[47734]: pgmap v343: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:44 vm10 ceph-mon[51861]: pgmap v343: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:45.881 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:45.883 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:45.923 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:45.924 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:46 vm08 ceph-mon[47734]: pgmap v344: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:04:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:04:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:04:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:04:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:04:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:04:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:04:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:46 vm10 ceph-mon[51861]: pgmap v344: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:04:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:04:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:04:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:04:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:04:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:04:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:04:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:48 vm08 ceph-mon[47734]: pgmap v345: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:48 vm10 ceph-mon[51861]: pgmap v345: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:50 vm08 ceph-mon[47734]: pgmap v346: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:50 vm10 ceph-mon[51861]: pgmap v346: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:50.925 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:50.926 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:50.951 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:50.951 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:52 vm10 ceph-mon[51861]: pgmap v347: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:52 vm08 ceph-mon[47734]: pgmap v347: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:54 vm10 ceph-mon[51861]: pgmap v348: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:54 vm08 ceph-mon[47734]: pgmap v348: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:55.953 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:04:55.953 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:04:55.978 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:04:55.978 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:04:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:56 vm08 ceph-mon[47734]: pgmap v349: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:56 vm10 ceph-mon[51861]: pgmap v349: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:04:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:04:58 vm10 ceph-mon[51861]: pgmap v350: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:04:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:04:58 vm08 ceph-mon[47734]: pgmap v350: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:00 vm10 ceph-mon[51861]: pgmap v351: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:00.980 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:00.980 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:01.012 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:01.012 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:00 vm08 ceph-mon[47734]: pgmap v351: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:01 vm10 ceph-mon[51861]: pgmap v352: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:01 vm08 ceph-mon[47734]: pgmap v352: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:04 vm10 ceph-mon[51861]: pgmap v353: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:04 vm08 ceph-mon[47734]: pgmap v353: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:06.014 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:06.014 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:06.100 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:06.101 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:06 vm08 ceph-mon[47734]: pgmap v354: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:06 vm10 ceph-mon[51861]: pgmap v354: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:08.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:08 vm08 ceph-mon[47734]: pgmap v355: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:08 vm10 ceph-mon[51861]: pgmap v355: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:10 vm10 ceph-mon[51861]: pgmap v356: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:10 vm08 ceph-mon[47734]: pgmap v356: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:11.103 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:11.103 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:11.129 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:11.130 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:12.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:12 vm10 ceph-mon[51861]: pgmap v357: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:12 vm08 ceph-mon[47734]: pgmap v357: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:14 vm10 ceph-mon[51861]: pgmap v358: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:14 vm08 ceph-mon[47734]: pgmap v358: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:16.131 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:16.132 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:16.158 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:16.159 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:16 vm08 ceph-mon[47734]: pgmap v359: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:16 vm10 ceph-mon[51861]: pgmap v359: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:18 vm10 ceph-mon[51861]: pgmap v360: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:18 vm08 ceph-mon[47734]: pgmap v360: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:20 vm10 ceph-mon[51861]: pgmap v361: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:20 vm08 ceph-mon[47734]: pgmap v361: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:21.160 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:21.161 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:21.188 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:21.189 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:22 vm10 ceph-mon[51861]: pgmap v362: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:22 vm08 ceph-mon[47734]: pgmap v362: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:24.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:23 vm08 ceph-mon[47734]: pgmap v363: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:24.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:23 vm10 ceph-mon[51861]: pgmap v363: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:26.191 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:26.191 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:26.219 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:26.219 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:26 vm08 ceph-mon[47734]: pgmap v364: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:26 vm10 ceph-mon[51861]: pgmap v364: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:28 vm10 ceph-mon[51861]: pgmap v365: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:28 vm08 ceph-mon[47734]: pgmap v365: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:30 vm10 ceph-mon[51861]: pgmap v366: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:30 vm08 ceph-mon[47734]: pgmap v366: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:31.221 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:31.221 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:31.247 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:31.247 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:32 vm10 ceph-mon[51861]: pgmap v367: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:32 vm08 ceph-mon[47734]: pgmap v367: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:34 vm10 ceph-mon[51861]: pgmap v368: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:34.957 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:34 vm08 ceph-mon[47734]: pgmap v368: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:36.249 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:36.249 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:36.275 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:36.275 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:36 vm08 ceph-mon[47734]: pgmap v369: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:36 vm10 ceph-mon[51861]: pgmap v369: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:38 vm10 ceph-mon[51861]: pgmap v370: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:38 vm08 ceph-mon[47734]: pgmap v370: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:40 vm10 ceph-mon[51861]: pgmap v371: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:40 vm08 ceph-mon[47734]: pgmap v371: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:41.276 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:41.277 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:41.303 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:41.304 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:42 vm10 ceph-mon[51861]: pgmap v372: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:42 vm08 ceph-mon[47734]: pgmap v372: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:05:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:05:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:05:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:05:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:44 vm10 ceph-mon[51861]: pgmap v373: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:44 vm08 ceph-mon[47734]: pgmap v373: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:46.306 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:46.307 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:46.349 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:46.350 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:46.646 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:46 vm10 ceph-mon[51861]: pgmap v374: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:46.646 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:05:46.646 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:05:46.646 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:05:46.646 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:05:46.647 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:05:46.647 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:46 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:05:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:46 vm08 ceph-mon[47734]: pgmap v374: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:05:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:05:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:05:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:05:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:05:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:46 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:05:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:48 vm10 ceph-mon[51861]: pgmap v375: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:48 vm08 ceph-mon[47734]: pgmap v375: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:50 vm10 ceph-mon[51861]: pgmap v376: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:50 vm08 ceph-mon[47734]: pgmap v376: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:51.352 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:51.353 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:51.378 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:51.378 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:52 vm08 ceph-mon[47734]: pgmap v377: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:52 vm10 ceph-mon[51861]: pgmap v377: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:54 vm08 ceph-mon[47734]: pgmap v378: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:54 vm10 ceph-mon[51861]: pgmap v378: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:56.380 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:05:56.380 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:05:56.407 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:05:56.408 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:05:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:56 vm08 ceph-mon[47734]: pgmap v379: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:56 vm10 ceph-mon[51861]: pgmap v379: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:05:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:05:58 vm08 ceph-mon[47734]: pgmap v380: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:05:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:05:58 vm10 ceph-mon[51861]: pgmap v380: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:00 vm08 ceph-mon[47734]: pgmap v381: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:00 vm10 ceph-mon[51861]: pgmap v381: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:01.409 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:01.410 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:01.436 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:01.436 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:02.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:02 vm08 ceph-mon[47734]: pgmap v382: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:02 vm10 ceph-mon[51861]: pgmap v382: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:04.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:04 vm08 ceph-mon[47734]: pgmap v383: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:04 vm10 ceph-mon[51861]: pgmap v383: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:06.438 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:06.438 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:06.465 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:06.465 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:06.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:06 vm08 ceph-mon[47734]: pgmap v384: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:06 vm10 ceph-mon[51861]: pgmap v384: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:08 vm10 ceph-mon[51861]: pgmap v385: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:08 vm08 ceph-mon[47734]: pgmap v385: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:10 vm10 ceph-mon[51861]: pgmap v386: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:10 vm08 ceph-mon[47734]: pgmap v386: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:11.467 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:11.468 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:11.495 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:11.496 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:12 vm10 ceph-mon[51861]: pgmap v387: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:12 vm08 ceph-mon[47734]: pgmap v387: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:14 vm10 ceph-mon[51861]: pgmap v388: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:14 vm08 ceph-mon[47734]: pgmap v388: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:16.497 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:16.498 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:16.524 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:16.524 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:16 vm08 ceph-mon[47734]: pgmap v389: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:16 vm10 ceph-mon[51861]: pgmap v389: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:18 vm10 ceph-mon[51861]: pgmap v390: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:18 vm08 ceph-mon[47734]: pgmap v390: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:20 vm10 ceph-mon[51861]: pgmap v391: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:20 vm08 ceph-mon[47734]: pgmap v391: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:21.526 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:21.526 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:21.552 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:21.553 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:22 vm10 ceph-mon[51861]: pgmap v392: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:22 vm08 ceph-mon[47734]: pgmap v392: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:24 vm10 ceph-mon[51861]: pgmap v393: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:24 vm08 ceph-mon[47734]: pgmap v393: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:26.554 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:26.554 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:26.580 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:26.581 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:26 vm08 ceph-mon[47734]: pgmap v394: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:26 vm10 ceph-mon[51861]: pgmap v394: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:28 vm10 ceph-mon[51861]: pgmap v395: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:28 vm08 ceph-mon[47734]: pgmap v395: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:30 vm10 ceph-mon[51861]: pgmap v396: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:30 vm08 ceph-mon[47734]: pgmap v396: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:31.582 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:31.583 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:31.608 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:31.609 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:32 vm10 ceph-mon[51861]: pgmap v397: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:32 vm08 ceph-mon[47734]: pgmap v397: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:34 vm10 ceph-mon[51861]: pgmap v398: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:34.958 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:34 vm08 ceph-mon[47734]: pgmap v398: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:36.610 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:36.611 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:36.636 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:36.636 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:36 vm08 ceph-mon[47734]: pgmap v399: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:36 vm10 ceph-mon[51861]: pgmap v399: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:38.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:38 vm10 ceph-mon[51861]: pgmap v400: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:38 vm08 ceph-mon[47734]: pgmap v400: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:40 vm08 ceph-mon[47734]: pgmap v401: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:40 vm10 ceph-mon[51861]: pgmap v401: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:41.638 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:41.638 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:41.665 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:41.666 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:42 vm08 ceph-mon[47734]: pgmap v402: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:42 vm10 ceph-mon[51861]: pgmap v402: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:06:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:06:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:43 vm08 ceph-mon[47734]: pgmap v403: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:06:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:06:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:43 vm10 ceph-mon[51861]: pgmap v403: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:46.667 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:46.668 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:46.702 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:46.702 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:46 vm08 ceph-mon[47734]: pgmap v404: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:46.854 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:46 vm10 ceph-mon[51861]: pgmap v404: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:47.770 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:47 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:06:47.770 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:47 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:06:47.770 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:47 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:06:47.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:47 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:06:47.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:47 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:06:47.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:47 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:06:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:48 vm08 ceph-mon[47734]: pgmap v405: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:48.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:06:48.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:06:48.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:06:48.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:06:48.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:06:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:48 vm10 ceph-mon[51861]: pgmap v405: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:48.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:06:48.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:06:48.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:06:48.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:06:48.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:06:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:50 vm08 ceph-mon[47734]: pgmap v406: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:50 vm10 ceph-mon[51861]: pgmap v406: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:51.704 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:51.704 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:51.734 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:51.734 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:52.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:52 vm08 ceph-mon[47734]: pgmap v407: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:52 vm10 ceph-mon[51861]: pgmap v407: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:54 vm08 ceph-mon[47734]: pgmap v408: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:54 vm10 ceph-mon[51861]: pgmap v408: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:56.736 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:06:56.736 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:06:56.762 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:06:56.763 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:06:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:56 vm08 ceph-mon[47734]: pgmap v409: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:56 vm10 ceph-mon[51861]: pgmap v409: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:06:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:06:58 vm10 ceph-mon[51861]: pgmap v410: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:06:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:06:58 vm08 ceph-mon[47734]: pgmap v410: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:00 vm10 ceph-mon[51861]: pgmap v411: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:00 vm08 ceph-mon[47734]: pgmap v411: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:01.764 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:01.765 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:01.790 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:01.791 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:02 vm10 ceph-mon[51861]: pgmap v412: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:02 vm08 ceph-mon[47734]: pgmap v412: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:04 vm10 ceph-mon[51861]: pgmap v413: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:04.958 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:04 vm08 ceph-mon[47734]: pgmap v413: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:06.793 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:06.793 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:06 vm08 ceph-mon[47734]: pgmap v414: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:06.820 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:06.821 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:06.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:06 vm10 ceph-mon[51861]: pgmap v414: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:08 vm10 ceph-mon[51861]: pgmap v415: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:08 vm08 ceph-mon[47734]: pgmap v415: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:10 vm10 ceph-mon[51861]: pgmap v416: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:10 vm08 ceph-mon[47734]: pgmap v416: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:11.822 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:11.823 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:11.848 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:11.848 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:12 vm10 ceph-mon[51861]: pgmap v417: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:12 vm08 ceph-mon[47734]: pgmap v417: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:14 vm10 ceph-mon[51861]: pgmap v418: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:14 vm08 ceph-mon[47734]: pgmap v418: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:16 vm08 ceph-mon[47734]: pgmap v419: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:16.849 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:16.850 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:16.877 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:16.878 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:16 vm10 ceph-mon[51861]: pgmap v419: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:18 vm10 ceph-mon[51861]: pgmap v420: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:18 vm08 ceph-mon[47734]: pgmap v420: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:20 vm10 ceph-mon[51861]: pgmap v421: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:20 vm08 ceph-mon[47734]: pgmap v421: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:21.879 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:21.880 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:21.905 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:21.905 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:22 vm10 ceph-mon[51861]: pgmap v422: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:22 vm08 ceph-mon[47734]: pgmap v422: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:24 vm10 ceph-mon[51861]: pgmap v423: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:24 vm08 ceph-mon[47734]: pgmap v423: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:26 vm08 ceph-mon[47734]: pgmap v424: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:26.907 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:26.907 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:26 vm10 ceph-mon[51861]: pgmap v424: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:26.933 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:26.933 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:28.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:28 vm10 ceph-mon[51861]: pgmap v425: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:28 vm08 ceph-mon[47734]: pgmap v425: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:30 vm08 ceph-mon[47734]: pgmap v426: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:30 vm10 ceph-mon[51861]: pgmap v426: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:31.935 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:31.935 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:31.990 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:31.991 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:32 vm08 ceph-mon[47734]: pgmap v427: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:32 vm10 ceph-mon[51861]: pgmap v427: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:34.958 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:34 vm08 ceph-mon[47734]: pgmap v428: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:34 vm10 ceph-mon[51861]: pgmap v428: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:36 vm08 ceph-mon[47734]: pgmap v429: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:36 vm10 ceph-mon[51861]: pgmap v429: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:36.992 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:36.992 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:37.017 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:37.018 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:38 vm08 ceph-mon[47734]: pgmap v430: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:38 vm10 ceph-mon[51861]: pgmap v430: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:40.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:40 vm08 ceph-mon[47734]: pgmap v431: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:40 vm10 ceph-mon[51861]: pgmap v431: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:42.019 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:42.020 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:42.045 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:42.046 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:42.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:42 vm08 ceph-mon[47734]: pgmap v432: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:42 vm10 ceph-mon[51861]: pgmap v432: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:07:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:07:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:07:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:07:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:44 vm08 ceph-mon[47734]: pgmap v433: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:44 vm10 ceph-mon[51861]: pgmap v433: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:46 vm08 ceph-mon[47734]: pgmap v434: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:46 vm10 ceph-mon[51861]: pgmap v434: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:47.047 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:47.048 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:47.073 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:47.074 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:48.609 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:48 vm10 ceph-mon[51861]: pgmap v435: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:48.609 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:07:48.609 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:07:48.609 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:07:48.609 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:07:48.609 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:07:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:48 vm08 ceph-mon[47734]: pgmap v435: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:49.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:07:49.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:07:49.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:07:49.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:07:49.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:07:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:50 vm10 ceph-mon[51861]: pgmap v436: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:50 vm08 ceph-mon[47734]: pgmap v436: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:52.075 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:52.076 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:52.101 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:52.102 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:52 vm08 ceph-mon[47734]: pgmap v437: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:52 vm10 ceph-mon[51861]: pgmap v437: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:54.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:53 vm08 ceph-mon[47734]: pgmap v438: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:54.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:53 vm10 ceph-mon[51861]: pgmap v438: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:56 vm08 ceph-mon[47734]: pgmap v439: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:56 vm10 ceph-mon[51861]: pgmap v439: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:07:57.103 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:07:57.104 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:07:57.129 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:07:57.129 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:07:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:07:58 vm08 ceph-mon[47734]: pgmap v440: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:07:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:07:58 vm10 ceph-mon[51861]: pgmap v440: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:00.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:00 vm08 ceph-mon[47734]: pgmap v441: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:00 vm10 ceph-mon[51861]: pgmap v441: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:02.130 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:02.131 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:02.158 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:02.158 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:02 vm10 ceph-mon[51861]: pgmap v442: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:02 vm08 ceph-mon[47734]: pgmap v442: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:04 vm10 ceph-mon[51861]: pgmap v443: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:04.957 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:04 vm08 ceph-mon[47734]: pgmap v443: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:06 vm08 ceph-mon[47734]: pgmap v444: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:06 vm10 ceph-mon[51861]: pgmap v444: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:07.159 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:07.160 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:07.186 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:07.187 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:08 vm10 ceph-mon[51861]: pgmap v445: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:08 vm08 ceph-mon[47734]: pgmap v445: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:10 vm10 ceph-mon[51861]: pgmap v446: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:10 vm08 ceph-mon[47734]: pgmap v446: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:12.188 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:12.188 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:12.216 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:12.217 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:12 vm10 ceph-mon[51861]: pgmap v447: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:12 vm08 ceph-mon[47734]: pgmap v447: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:14 vm10 ceph-mon[51861]: pgmap v448: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:14 vm08 ceph-mon[47734]: pgmap v448: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:16 vm08 ceph-mon[47734]: pgmap v449: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:16 vm10 ceph-mon[51861]: pgmap v449: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:17.218 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:17.219 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:17.245 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:17.245 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:18 vm10 ceph-mon[51861]: pgmap v450: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:18 vm08 ceph-mon[47734]: pgmap v450: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:20 vm08 ceph-mon[47734]: pgmap v451: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:20 vm10 ceph-mon[51861]: pgmap v451: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:22.247 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:22.247 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:22.273 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:22.273 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:22 vm08 ceph-mon[47734]: pgmap v452: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:22 vm10 ceph-mon[51861]: pgmap v452: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:24 vm08 ceph-mon[47734]: pgmap v453: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:24 vm10 ceph-mon[51861]: pgmap v453: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:26 vm08 ceph-mon[47734]: pgmap v454: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:26 vm10 ceph-mon[51861]: pgmap v454: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:27.274 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:27.275 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:27.300 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:27.300 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:28.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:27 vm08 ceph-mon[47734]: pgmap v455: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:28.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:27 vm10 ceph-mon[51861]: pgmap v455: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:30 vm10 ceph-mon[51861]: pgmap v456: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:30 vm08 ceph-mon[47734]: pgmap v456: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:32.301 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:32.302 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:32.331 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:32.331 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:32 vm08 ceph-mon[47734]: pgmap v457: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:32 vm10 ceph-mon[51861]: pgmap v457: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:34 vm08 ceph-mon[47734]: pgmap v458: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:34 vm10 ceph-mon[51861]: pgmap v458: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:36 vm08 ceph-mon[47734]: pgmap v459: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:36 vm10 ceph-mon[51861]: pgmap v459: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:37.333 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:37.333 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:37.360 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:37.360 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:38 vm10 ceph-mon[51861]: pgmap v460: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:38 vm08 ceph-mon[47734]: pgmap v460: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:40 vm10 ceph-mon[51861]: pgmap v461: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:40 vm08 ceph-mon[47734]: pgmap v461: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:42.362 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:42.362 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:42.389 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:42.390 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:42 vm10 ceph-mon[51861]: pgmap v462: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:42 vm08 ceph-mon[47734]: pgmap v462: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:08:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:08:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:08:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:08:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:44 vm10 ceph-mon[51861]: pgmap v463: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:44 vm08 ceph-mon[47734]: pgmap v463: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:46 vm08 ceph-mon[47734]: pgmap v464: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:46 vm10 ceph-mon[51861]: pgmap v464: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:47.391 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:47.392 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:47.420 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:47.420 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:48 vm10 ceph-mon[51861]: pgmap v465: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:08:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:08:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:48 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:08:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:48 vm08 ceph-mon[47734]: pgmap v465: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:08:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:08:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:48 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:08:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:08:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:08:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:49 vm08 ceph-mon[47734]: pgmap v466: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:08:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:08:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:49 vm10 ceph-mon[51861]: pgmap v466: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:52.422 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:52.422 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:52.451 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:52.452 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:52 vm08 ceph-mon[47734]: pgmap v467: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:52 vm10 ceph-mon[51861]: pgmap v467: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:54 vm08 ceph-mon[47734]: pgmap v468: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:54 vm10 ceph-mon[51861]: pgmap v468: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:56 vm08 ceph-mon[47734]: pgmap v469: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:56 vm10 ceph-mon[51861]: pgmap v469: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:08:57.453 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:08:57.453 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:08:57.479 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:08:57.480 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:08:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:58 vm08 ceph-mon[47734]: pgmap v470: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:08:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:58 vm10 ceph-mon[51861]: pgmap v470: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:00.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:08:59 vm10 ceph-mon[51861]: pgmap v471: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:00.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:08:59 vm08 ceph-mon[47734]: pgmap v471: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:02.481 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:02.482 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:02.507 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:02.508 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:02.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:02 vm08 ceph-mon[47734]: pgmap v472: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:02 vm10 ceph-mon[51861]: pgmap v472: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:04 vm08 ceph-mon[47734]: pgmap v473: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:04 vm10 ceph-mon[51861]: pgmap v473: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:06 vm08 ceph-mon[47734]: pgmap v474: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:06 vm10 ceph-mon[51861]: pgmap v474: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:07.509 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:07.510 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:07.536 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:07.536 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:08.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:08 vm08 ceph-mon[47734]: pgmap v475: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:08 vm10 ceph-mon[51861]: pgmap v475: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:10.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:10 vm08 ceph-mon[47734]: pgmap v476: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:10 vm10 ceph-mon[51861]: pgmap v476: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:12.538 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:12.538 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:12.568 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:12.568 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:12.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:12 vm08 ceph-mon[47734]: pgmap v477: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:12 vm10 ceph-mon[51861]: pgmap v477: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:14 vm08 ceph-mon[47734]: pgmap v478: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:14 vm10 ceph-mon[51861]: pgmap v478: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:16 vm08 ceph-mon[47734]: pgmap v479: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:16 vm10 ceph-mon[51861]: pgmap v479: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:17.570 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:17.570 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:17.597 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:17.597 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:18 vm08 ceph-mon[47734]: pgmap v480: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:18 vm10 ceph-mon[51861]: pgmap v480: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:20 vm10 ceph-mon[51861]: pgmap v481: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:20 vm08 ceph-mon[47734]: pgmap v481: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:22.599 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:22.599 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:22.624 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:22.625 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:22 vm10 ceph-mon[51861]: pgmap v482: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:22 vm08 ceph-mon[47734]: pgmap v482: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:24 vm10 ceph-mon[51861]: pgmap v483: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:24 vm08 ceph-mon[47734]: pgmap v483: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:26 vm08 ceph-mon[47734]: pgmap v484: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:26 vm10 ceph-mon[51861]: pgmap v484: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:27.626 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:27.626 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:27.651 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:27.651 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:28 vm10 ceph-mon[51861]: pgmap v485: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:28 vm08 ceph-mon[47734]: pgmap v485: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:30 vm10 ceph-mon[51861]: pgmap v486: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:30 vm08 ceph-mon[47734]: pgmap v486: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:32.653 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:32.653 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:32.679 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:32.680 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:32 vm10 ceph-mon[51861]: pgmap v487: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:32 vm08 ceph-mon[47734]: pgmap v487: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:34 vm10 ceph-mon[51861]: pgmap v488: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:34.958 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:34 vm08 ceph-mon[47734]: pgmap v488: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:36 vm08 ceph-mon[47734]: pgmap v489: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:36 vm10 ceph-mon[51861]: pgmap v489: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:37.681 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:37.682 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:37.707 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:37.708 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:38 vm10 ceph-mon[51861]: pgmap v490: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:38 vm08 ceph-mon[47734]: pgmap v490: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:40 vm10 ceph-mon[51861]: pgmap v491: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:40 vm08 ceph-mon[47734]: pgmap v491: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:42.709 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:42.709 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:42.738 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:42.739 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:42 vm10 ceph-mon[51861]: pgmap v492: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:42 vm08 ceph-mon[47734]: pgmap v492: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:09:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:09:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:09:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:09:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:44 vm10 ceph-mon[51861]: pgmap v493: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:44 vm08 ceph-mon[47734]: pgmap v493: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:46 vm08 ceph-mon[47734]: pgmap v494: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:46 vm10 ceph-mon[51861]: pgmap v494: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:47.740 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:47.741 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:47.766 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:47.767 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:48 vm10 ceph-mon[51861]: pgmap v495: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:49.018 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:48 vm08 ceph-mon[47734]: pgmap v495: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:09:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:09:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:09:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:09:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:09:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:09:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:09:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:09:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:09:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:09:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:50 vm10 ceph-mon[51861]: pgmap v496: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:50 vm08 ceph-mon[47734]: pgmap v496: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:52.768 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:52.768 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:52.795 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:52.795 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:52 vm10 ceph-mon[51861]: pgmap v497: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:52 vm08 ceph-mon[47734]: pgmap v497: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:54 vm10 ceph-mon[51861]: pgmap v498: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:54 vm08 ceph-mon[47734]: pgmap v498: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:56 vm08 ceph-mon[47734]: pgmap v499: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:56 vm10 ceph-mon[51861]: pgmap v499: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:09:57.796 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:09:57.797 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:09:57.823 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:09:57.823 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:09:58.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:09:58 vm10 ceph-mon[51861]: pgmap v500: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:09:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:09:58 vm08 ceph-mon[47734]: pgmap v500: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:00 vm08 ceph-mon[47734]: pgmap v501: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:00 vm08 ceph-mon[47734]: overall HEALTH_OK 2026-03-09T14:10:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:00 vm10 ceph-mon[51861]: pgmap v501: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:00 vm10 ceph-mon[51861]: overall HEALTH_OK 2026-03-09T14:10:02.825 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:02.825 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:02.850 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:02.851 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:02 vm08 ceph-mon[47734]: pgmap v502: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:02 vm10 ceph-mon[51861]: pgmap v502: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:04.957 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:04 vm08 ceph-mon[47734]: pgmap v503: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:04 vm10 ceph-mon[51861]: pgmap v503: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:06.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:06 vm08 ceph-mon[47734]: pgmap v504: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:06 vm10 ceph-mon[51861]: pgmap v504: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:07.852 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:07.853 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:07.877 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:07.878 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:08.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:08 vm10 ceph-mon[51861]: pgmap v505: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:08 vm08 ceph-mon[47734]: pgmap v505: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:10 vm08 ceph-mon[47734]: pgmap v506: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:10 vm10 ceph-mon[51861]: pgmap v506: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:12.879 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:12.880 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:12.905 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:12.906 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:12 vm08 ceph-mon[47734]: pgmap v507: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:12 vm10 ceph-mon[51861]: pgmap v507: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:14.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:13 vm08 ceph-mon[47734]: pgmap v508: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:14.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:13 vm10 ceph-mon[51861]: pgmap v508: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:16 vm08 ceph-mon[47734]: pgmap v509: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:16 vm10 ceph-mon[51861]: pgmap v509: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:17.907 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:17.907 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:17.939 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:17.940 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:18 vm08 ceph-mon[47734]: pgmap v510: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:18 vm10 ceph-mon[51861]: pgmap v510: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:20 vm10 ceph-mon[51861]: pgmap v511: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:20 vm08 ceph-mon[47734]: pgmap v511: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:22 vm10 ceph-mon[51861]: pgmap v512: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:22.941 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:22.942 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:22.967 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:22.967 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:22 vm08 ceph-mon[47734]: pgmap v512: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:24 vm10 ceph-mon[51861]: pgmap v513: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:24 vm08 ceph-mon[47734]: pgmap v513: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:26 vm08 ceph-mon[47734]: pgmap v514: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:26 vm10 ceph-mon[51861]: pgmap v514: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:27.969 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:27.969 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:28.037 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:28.038 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:28 vm10 ceph-mon[51861]: pgmap v515: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:28 vm08 ceph-mon[47734]: pgmap v515: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:30 vm10 ceph-mon[51861]: pgmap v516: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:30 vm08 ceph-mon[47734]: pgmap v516: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:32 vm10 ceph-mon[51861]: pgmap v517: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:33.039 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:33.040 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:32 vm08 ceph-mon[47734]: pgmap v517: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:33.065 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:33.066 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:34 vm10 ceph-mon[51861]: pgmap v518: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:34.957 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:34 vm08 ceph-mon[47734]: pgmap v518: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:36 vm08 ceph-mon[47734]: pgmap v519: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:36 vm10 ceph-mon[51861]: pgmap v519: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:38.067 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:38.068 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:38.094 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:38.094 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:38.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:38 vm10 ceph-mon[51861]: pgmap v520: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:38 vm08 ceph-mon[47734]: pgmap v520: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:40 vm08 ceph-mon[47734]: pgmap v521: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:40 vm10 ceph-mon[51861]: pgmap v521: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:42 vm08 ceph-mon[47734]: pgmap v522: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:43.096 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:43.096 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:43.122 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:43.122 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:42 vm10 ceph-mon[51861]: pgmap v522: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:10:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:10:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:10:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:10:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:44 vm08 ceph-mon[47734]: pgmap v523: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:44 vm10 ceph-mon[51861]: pgmap v523: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:46.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:46 vm08 ceph-mon[47734]: pgmap v524: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:46 vm10 ceph-mon[51861]: pgmap v524: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:48.123 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:48.124 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:48.150 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:48.151 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:48 vm08 ceph-mon[47734]: pgmap v525: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:49.060 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:48 vm10 ceph-mon[51861]: pgmap v525: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:10:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:10:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:10:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:49 vm08 ceph-mon[47734]: pgmap v526: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:10:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:10:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:10:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:10:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:10:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:49 vm10 ceph-mon[51861]: pgmap v526: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:10:50.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:10:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:52 vm08 ceph-mon[47734]: pgmap v527: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:52 vm10 ceph-mon[51861]: pgmap v527: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:53.152 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:53.153 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:53.179 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:53.179 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:54 vm08 ceph-mon[47734]: pgmap v528: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:54 vm10 ceph-mon[51861]: pgmap v528: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:56 vm08 ceph-mon[47734]: pgmap v529: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:56 vm10 ceph-mon[51861]: pgmap v529: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:10:58.181 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:10:58.181 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:10:58.228 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:10:58.228 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:10:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:10:58 vm10 ceph-mon[51861]: pgmap v530: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:10:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:10:58 vm08 ceph-mon[47734]: pgmap v530: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:00 vm10 ceph-mon[51861]: pgmap v531: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:00 vm08 ceph-mon[47734]: pgmap v531: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:02 vm10 ceph-mon[51861]: pgmap v532: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:02 vm08 ceph-mon[47734]: pgmap v532: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:03.229 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:03.230 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:03.259 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:03.259 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:04 vm10 ceph-mon[51861]: pgmap v533: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:04.957 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:04 vm08 ceph-mon[47734]: pgmap v533: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:06.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:06 vm08 ceph-mon[47734]: pgmap v534: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:06 vm10 ceph-mon[51861]: pgmap v534: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:08.261 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:08.433 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:08.521 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:08.521 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:08 vm10 ceph-mon[51861]: pgmap v535: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:08 vm08 ceph-mon[47734]: pgmap v535: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:10 vm10 ceph-mon[51861]: pgmap v536: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:10 vm08 ceph-mon[47734]: pgmap v536: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:12 vm10 ceph-mon[51861]: pgmap v537: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:12 vm08 ceph-mon[47734]: pgmap v537: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:13.523 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:13.523 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:13.551 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:13.551 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:14 vm10 ceph-mon[51861]: pgmap v538: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:14 vm08 ceph-mon[47734]: pgmap v538: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:16 vm08 ceph-mon[47734]: pgmap v539: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:16 vm10 ceph-mon[51861]: pgmap v539: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:18.553 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:18.553 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:18.579 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:18.580 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:18 vm10 ceph-mon[51861]: pgmap v540: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:18 vm08 ceph-mon[47734]: pgmap v540: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:20 vm10 ceph-mon[51861]: pgmap v541: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:20 vm08 ceph-mon[47734]: pgmap v541: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:22 vm10 ceph-mon[51861]: pgmap v542: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:22 vm08 ceph-mon[47734]: pgmap v542: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:23.581 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:23.582 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:23.608 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:23.609 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:24 vm08 ceph-mon[47734]: pgmap v543: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:24 vm10 ceph-mon[51861]: pgmap v543: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:26.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:26 vm08 ceph-mon[47734]: pgmap v544: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:26 vm10 ceph-mon[51861]: pgmap v544: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:28.610 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:28.611 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:28.637 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:28.638 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:28.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:28 vm10 ceph-mon[51861]: pgmap v545: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:28 vm08 ceph-mon[47734]: pgmap v545: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:30 vm08 ceph-mon[47734]: pgmap v546: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:30 vm10 ceph-mon[51861]: pgmap v546: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:32.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:31 vm08 ceph-mon[47734]: pgmap v547: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:32.059 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:31 vm10 ceph-mon[51861]: pgmap v547: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:33.639 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:33.640 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:33.666 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:33.667 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:34.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:34 vm10 ceph-mon[51861]: pgmap v548: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:34 vm08 ceph-mon[47734]: pgmap v548: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:36 vm08 ceph-mon[47734]: pgmap v549: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:36 vm10 ceph-mon[51861]: pgmap v549: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:38.668 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:38.669 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:38.694 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:38.694 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:38 vm10 ceph-mon[51861]: pgmap v550: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:38 vm08 ceph-mon[47734]: pgmap v550: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:40 vm10 ceph-mon[51861]: pgmap v551: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:40 vm08 ceph-mon[47734]: pgmap v551: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:42 vm10 ceph-mon[51861]: pgmap v552: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:11:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:42 vm08 ceph-mon[47734]: pgmap v552: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:11:43.696 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:43.696 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:43.721 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:43.722 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:11:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:11:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:11:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:11:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:44 vm08 ceph-mon[47734]: pgmap v553: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:11:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:44 vm10 ceph-mon[51861]: pgmap v553: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:11:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:46 vm08 ceph-mon[47734]: pgmap v554: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:46 vm10 ceph-mon[51861]: pgmap v554: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:48.723 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:48.724 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:48.750 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:48.750 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:48 vm10 ceph-mon[51861]: pgmap v555: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:11:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:48 vm08 ceph-mon[47734]: pgmap v555: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:11:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:11:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:11:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:11:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:11:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:11:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:11:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:11:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:11:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:11:50.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:11:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:50 vm10 ceph-mon[51861]: pgmap v556: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:50 vm08 ceph-mon[47734]: pgmap v556: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:52 vm10 ceph-mon[51861]: pgmap v557: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:52 vm08 ceph-mon[47734]: pgmap v557: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:53.752 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:53.752 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:53.779 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:53.779 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:54 vm10 ceph-mon[51861]: pgmap v558: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:54 vm08 ceph-mon[47734]: pgmap v558: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:56 vm08 ceph-mon[47734]: pgmap v559: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:56 vm10 ceph-mon[51861]: pgmap v559: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:11:58.780 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:11:58.781 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:11:58.807 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:11:58.807 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:11:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:11:58 vm10 ceph-mon[51861]: pgmap v560: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:11:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:11:58 vm08 ceph-mon[47734]: pgmap v560: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:00 vm10 ceph-mon[51861]: pgmap v561: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:00 vm08 ceph-mon[47734]: pgmap v561: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:02 vm10 ceph-mon[51861]: pgmap v562: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:02 vm08 ceph-mon[47734]: pgmap v562: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:03.809 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:03.809 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:03.835 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:03.835 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:04 vm10 ceph-mon[51861]: pgmap v563: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:04.958 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:04 vm08 ceph-mon[47734]: pgmap v563: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:06 vm08 ceph-mon[47734]: pgmap v564: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:06 vm10 ceph-mon[51861]: pgmap v564: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:08.837 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:08.837 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:08.863 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:08.863 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:08 vm10 ceph-mon[51861]: pgmap v565: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:08 vm08 ceph-mon[47734]: pgmap v565: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:10 vm10 ceph-mon[51861]: pgmap v566: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:10 vm08 ceph-mon[47734]: pgmap v566: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:11 vm10 ceph-mon[51861]: pgmap v567: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:11 vm08 ceph-mon[47734]: pgmap v567: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:13.865 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:13.865 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:13.895 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:13.896 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:14 vm08 ceph-mon[47734]: pgmap v568: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:14 vm10 ceph-mon[51861]: pgmap v568: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:16 vm08 ceph-mon[47734]: pgmap v569: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:16 vm10 ceph-mon[51861]: pgmap v569: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:18.897 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:18.898 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:18 vm10 ceph-mon[51861]: pgmap v570: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:18.923 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:18.923 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:18 vm08 ceph-mon[47734]: pgmap v570: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:20 vm10 ceph-mon[51861]: pgmap v571: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:20 vm08 ceph-mon[47734]: pgmap v571: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:22 vm10 ceph-mon[51861]: pgmap v572: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:22 vm08 ceph-mon[47734]: pgmap v572: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:23.925 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:23.925 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:23.952 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:23.953 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:24 vm10 ceph-mon[51861]: pgmap v573: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:24 vm08 ceph-mon[47734]: pgmap v573: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:26 vm08 ceph-mon[47734]: pgmap v574: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:26 vm10 ceph-mon[51861]: pgmap v574: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:28.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:28 vm10 ceph-mon[51861]: pgmap v575: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:28.954 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:28.954 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:28.986 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:28.986 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:28 vm08 ceph-mon[47734]: pgmap v575: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:30 vm08 ceph-mon[47734]: pgmap v576: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:30 vm10 ceph-mon[51861]: pgmap v576: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:32 vm08 ceph-mon[47734]: pgmap v577: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:32 vm10 ceph-mon[51861]: pgmap v577: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:33.987 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:33.988 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:34.015 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:34.015 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:34.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:33 vm08 ceph-mon[47734]: pgmap v578: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:34.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:33 vm10 ceph-mon[51861]: pgmap v578: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:36.760 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:36 vm08 ceph-mon[47734]: pgmap v579: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:36 vm10 ceph-mon[51861]: pgmap v579: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:38 vm10 ceph-mon[51861]: pgmap v580: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:39.017 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:39.017 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:39.044 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:39.044 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:38 vm08 ceph-mon[47734]: pgmap v580: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:40 vm10 ceph-mon[51861]: pgmap v581: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:40 vm08 ceph-mon[47734]: pgmap v581: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:42 vm08 ceph-mon[47734]: pgmap v582: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:42 vm10 ceph-mon[51861]: pgmap v582: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:44.046 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:44.046 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:12:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:12:44.074 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:44.074 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:12:44.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:12:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:44 vm08 ceph-mon[47734]: pgmap v583: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:44 vm10 ceph-mon[51861]: pgmap v583: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:46 vm08 ceph-mon[47734]: pgmap v584: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:46 vm10 ceph-mon[51861]: pgmap v584: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:48 vm08 ceph-mon[47734]: pgmap v585: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:49.075 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:49.076 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:49.101 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:49.102 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:48 vm10 ceph-mon[51861]: pgmap v585: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:49.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:12:49.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:12:49.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:49 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:12:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:12:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:12:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:49 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:12:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:50 vm08 ceph-mon[47734]: pgmap v586: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:12:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:12:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:50 vm10 ceph-mon[51861]: pgmap v586: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:12:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:12:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:52 vm08 ceph-mon[47734]: pgmap v587: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:52 vm10 ceph-mon[51861]: pgmap v587: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:54.103 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:54.104 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:54.129 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:54.129 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:12:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:54 vm08 ceph-mon[47734]: pgmap v588: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:54 vm10 ceph-mon[51861]: pgmap v588: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:56 vm08 ceph-mon[47734]: pgmap v589: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:56 vm10 ceph-mon[51861]: pgmap v589: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:12:58.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:12:57 vm10 ceph-mon[51861]: pgmap v590: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:12:57 vm08 ceph-mon[47734]: pgmap v590: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:12:59.130 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:12:59.131 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:12:59.155 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:12:59.155 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:00 vm08 ceph-mon[47734]: pgmap v591: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:00 vm10 ceph-mon[51861]: pgmap v591: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:02 vm10 ceph-mon[51861]: pgmap v592: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:02 vm08 ceph-mon[47734]: pgmap v592: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:04.156 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:04.157 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:04.184 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:04.185 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:04 vm10 ceph-mon[51861]: pgmap v593: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:04 vm08 ceph-mon[47734]: pgmap v593: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:06 vm08 ceph-mon[47734]: pgmap v594: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:06 vm10 ceph-mon[51861]: pgmap v594: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:08 vm10 ceph-mon[51861]: pgmap v595: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:08 vm08 ceph-mon[47734]: pgmap v595: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:09.186 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:09.187 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:09.212 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:09.212 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:10 vm08 ceph-mon[47734]: pgmap v596: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:10 vm10 ceph-mon[51861]: pgmap v596: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:12 vm08 ceph-mon[47734]: pgmap v597: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:12 vm10 ceph-mon[51861]: pgmap v597: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:14.214 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:14.214 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:14.244 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:14.244 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:14 vm08 ceph-mon[47734]: pgmap v598: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:14 vm10 ceph-mon[51861]: pgmap v598: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:16 vm08 ceph-mon[47734]: pgmap v599: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:16 vm10 ceph-mon[51861]: pgmap v599: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:18.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:18 vm10 ceph-mon[51861]: pgmap v600: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:18 vm08 ceph-mon[47734]: pgmap v600: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:19.245 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:19.246 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:19.271 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:19.272 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:20 vm08 ceph-mon[47734]: pgmap v601: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:20 vm10 ceph-mon[51861]: pgmap v601: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:22 vm08 ceph-mon[47734]: pgmap v602: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:22 vm10 ceph-mon[51861]: pgmap v602: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:24.273 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:24.274 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:24.300 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:24.301 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:24 vm08 ceph-mon[47734]: pgmap v603: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:24 vm10 ceph-mon[51861]: pgmap v603: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:26 vm08 ceph-mon[47734]: pgmap v604: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:26 vm10 ceph-mon[51861]: pgmap v604: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:28 vm08 ceph-mon[47734]: pgmap v605: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:28 vm10 ceph-mon[51861]: pgmap v605: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:29.302 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:29.302 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:29.327 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:29.327 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:30 vm08 ceph-mon[47734]: pgmap v606: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:30 vm10 ceph-mon[51861]: pgmap v606: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:32 vm08 ceph-mon[47734]: pgmap v607: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:33.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:32 vm10 ceph-mon[51861]: pgmap v607: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:34.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:33 vm08 ceph-mon[47734]: pgmap v608: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:34.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:33 vm10 ceph-mon[51861]: pgmap v608: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:34.329 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:34.329 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:34.356 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:34.356 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:36.759 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:36 vm08 ceph-mon[47734]: pgmap v609: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:36 vm10 ceph-mon[51861]: pgmap v609: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:38.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:38 vm10 ceph-mon[51861]: pgmap v610: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:38 vm08 ceph-mon[47734]: pgmap v610: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:39.358 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:39.358 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:39.384 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:39.385 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:40 vm08 ceph-mon[47734]: pgmap v611: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:40 vm10 ceph-mon[51861]: pgmap v611: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:42 vm08 ceph-mon[47734]: pgmap v612: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:42 vm10 ceph-mon[51861]: pgmap v612: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:13:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:13:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:13:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:13:44.386 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:44.387 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:44.412 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:44.412 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:44 vm08 ceph-mon[47734]: pgmap v613: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:44 vm10 ceph-mon[51861]: pgmap v613: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:46 vm08 ceph-mon[47734]: pgmap v614: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:46 vm10 ceph-mon[51861]: pgmap v614: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:48.940 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:48 vm10 ceph-mon[51861]: pgmap v615: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:48 vm08 ceph-mon[47734]: pgmap v615: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:49.413 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:49.414 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:49.439 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:49.439 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:50 vm08 ceph-mon[47734]: pgmap v616: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:13:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:13:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:13:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:13:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:13:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:50 vm10 ceph-mon[51861]: pgmap v616: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:13:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:13:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:13:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:13:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:13:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:52 vm08 ceph-mon[47734]: pgmap v617: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:52 vm10 ceph-mon[51861]: pgmap v617: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:54.441 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:54.441 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:54.467 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:54.467 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:13:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:54 vm08 ceph-mon[47734]: pgmap v618: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:54 vm10 ceph-mon[51861]: pgmap v618: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:56 vm08 ceph-mon[47734]: pgmap v619: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:56 vm10 ceph-mon[51861]: pgmap v619: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:13:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:13:58 vm08 ceph-mon[47734]: pgmap v620: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:13:58 vm10 ceph-mon[51861]: pgmap v620: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:13:59.469 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:13:59.469 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:13:59.495 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:13:59.495 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:00 vm08 ceph-mon[47734]: pgmap v621: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:00 vm10 ceph-mon[51861]: pgmap v621: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:02 vm08 ceph-mon[47734]: pgmap v622: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:02 vm10 ceph-mon[51861]: pgmap v622: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:03 vm08 ceph-mon[47734]: pgmap v623: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:04.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:03 vm10 ceph-mon[51861]: pgmap v623: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:04.497 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:04.497 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:04.522 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:04.523 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:06.759 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:06 vm08 ceph-mon[47734]: pgmap v624: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:06 vm10 ceph-mon[51861]: pgmap v624: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:08.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:08 vm10 ceph-mon[51861]: pgmap v625: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:08 vm08 ceph-mon[47734]: pgmap v625: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:09.524 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:09.525 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:09.550 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:09.551 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:10 vm08 ceph-mon[47734]: pgmap v626: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:10 vm10 ceph-mon[51861]: pgmap v626: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:12 vm08 ceph-mon[47734]: pgmap v627: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:12 vm10 ceph-mon[51861]: pgmap v627: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:14.552 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:14.553 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:14.577 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:14.578 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:14 vm08 ceph-mon[47734]: pgmap v628: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:14 vm10 ceph-mon[51861]: pgmap v628: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:16 vm08 ceph-mon[47734]: pgmap v629: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:16 vm10 ceph-mon[51861]: pgmap v629: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:18 vm08 ceph-mon[47734]: pgmap v630: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:18 vm10 ceph-mon[51861]: pgmap v630: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:19.579 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:19.580 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:19.608 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:19.609 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:20 vm08 ceph-mon[47734]: pgmap v631: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:20 vm10 ceph-mon[51861]: pgmap v631: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:22 vm08 ceph-mon[47734]: pgmap v632: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:22 vm10 ceph-mon[51861]: pgmap v632: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:24.610 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:24.611 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:24.636 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:24.636 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:24 vm08 ceph-mon[47734]: pgmap v633: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:24 vm10 ceph-mon[51861]: pgmap v633: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:27.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:26 vm08 ceph-mon[47734]: pgmap v634: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:26 vm10 ceph-mon[51861]: pgmap v634: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:28.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:27 vm08 ceph-mon[47734]: pgmap v635: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:28.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:27 vm10 ceph-mon[51861]: pgmap v635: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:29.638 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:29.638 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:29.694 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:29.694 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:30 vm08 ceph-mon[47734]: pgmap v636: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:30 vm10 ceph-mon[51861]: pgmap v636: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:32 vm08 ceph-mon[47734]: pgmap v637: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:32 vm10 ceph-mon[51861]: pgmap v637: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:34.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:33 vm08 ceph-mon[47734]: pgmap v638: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:34.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:33 vm10 ceph-mon[51861]: pgmap v638: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:34.696 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:34.696 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:34.722 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:34.723 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:36.759 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:36 vm08 ceph-mon[47734]: pgmap v639: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:36 vm10 ceph-mon[51861]: pgmap v639: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:38.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:38 vm10 ceph-mon[51861]: pgmap v640: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:38 vm08 ceph-mon[47734]: pgmap v640: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:39.724 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:39.725 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:39.753 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:39.754 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:40 vm08 ceph-mon[47734]: pgmap v641: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:40 vm10 ceph-mon[51861]: pgmap v641: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:42 vm08 ceph-mon[47734]: pgmap v642: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:42 vm10 ceph-mon[51861]: pgmap v642: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:14:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:14:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:14:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:14:44.755 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:44.756 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:44.782 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:44.782 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:44 vm08 ceph-mon[47734]: pgmap v643: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:44 vm10 ceph-mon[51861]: pgmap v643: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:46 vm08 ceph-mon[47734]: pgmap v644: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:46 vm10 ceph-mon[51861]: pgmap v644: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:48 vm08 ceph-mon[47734]: pgmap v645: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:48 vm10 ceph-mon[51861]: pgmap v645: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:49.784 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:49.785 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:49.810 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:49.811 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:50 vm08 ceph-mon[47734]: pgmap v646: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:14:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:14:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:14:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:14:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:14:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:14:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:50 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:14:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:50 vm10 ceph-mon[51861]: pgmap v646: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:14:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:14:51.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:14:51.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:14:51.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:14:51.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:14:51.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:50 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:14:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:52 vm08 ceph-mon[47734]: pgmap v647: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:52 vm10 ceph-mon[51861]: pgmap v647: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:54.812 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:54.813 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:54.840 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:54.841 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:14:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:54 vm08 ceph-mon[47734]: pgmap v648: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:54 vm10 ceph-mon[51861]: pgmap v648: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:56 vm08 ceph-mon[47734]: pgmap v649: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:56 vm10 ceph-mon[51861]: pgmap v649: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:14:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:14:58 vm08 ceph-mon[47734]: pgmap v650: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:14:58 vm10 ceph-mon[51861]: pgmap v650: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:14:59.842 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:14:59.843 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:14:59.918 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:14:59.918 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:00 vm08 ceph-mon[47734]: pgmap v651: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:00 vm10 ceph-mon[51861]: pgmap v651: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:02.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:01 vm08 ceph-mon[47734]: pgmap v652: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:01 vm10 ceph-mon[51861]: pgmap v652: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:04 vm08 ceph-mon[47734]: pgmap v653: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:04 vm10 ceph-mon[51861]: pgmap v653: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:04.920 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:04.921 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:04.947 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:04.948 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:06.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:06 vm08 ceph-mon[47734]: pgmap v654: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:06 vm10 ceph-mon[51861]: pgmap v654: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:08.940 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:08 vm10 ceph-mon[51861]: pgmap v655: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:08 vm08 ceph-mon[47734]: pgmap v655: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:09.950 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:09.950 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:09.979 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:09.980 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:10 vm08 ceph-mon[47734]: pgmap v656: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:10 vm10 ceph-mon[51861]: pgmap v656: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:12 vm08 ceph-mon[47734]: pgmap v657: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:12 vm10 ceph-mon[51861]: pgmap v657: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:14.981 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:14.981 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:15.010 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:15.010 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:14 vm08 ceph-mon[47734]: pgmap v658: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:14 vm10 ceph-mon[51861]: pgmap v658: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:16 vm08 ceph-mon[47734]: pgmap v659: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:16 vm10 ceph-mon[51861]: pgmap v659: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:18 vm08 ceph-mon[47734]: pgmap v660: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:18 vm10 ceph-mon[51861]: pgmap v660: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:20.012 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:20.012 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:20.041 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:20.042 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:20 vm08 ceph-mon[47734]: pgmap v661: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:20 vm10 ceph-mon[51861]: pgmap v661: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:22 vm08 ceph-mon[47734]: pgmap v662: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:22 vm10 ceph-mon[51861]: pgmap v662: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:25.043 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:25.044 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:24 vm08 ceph-mon[47734]: pgmap v663: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:25.069 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:25.069 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:24 vm10 ceph-mon[51861]: pgmap v663: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:26 vm08 ceph-mon[47734]: pgmap v664: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:26 vm10 ceph-mon[51861]: pgmap v664: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:28 vm08 ceph-mon[47734]: pgmap v665: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:28 vm10 ceph-mon[51861]: pgmap v665: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:30.071 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:30.071 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:30.096 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:30.097 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:30 vm08 ceph-mon[47734]: pgmap v666: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:30 vm10 ceph-mon[51861]: pgmap v666: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:32 vm08 ceph-mon[47734]: pgmap v667: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:32 vm10 ceph-mon[51861]: pgmap v667: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:34.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:33 vm08 ceph-mon[47734]: pgmap v668: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:34.059 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:33 vm10 ceph-mon[51861]: pgmap v668: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:35.098 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:35.099 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:35.124 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:35.124 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:36.760 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:36 vm08 ceph-mon[47734]: pgmap v669: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:36 vm10 ceph-mon[51861]: pgmap v669: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:38 vm08 ceph-mon[47734]: pgmap v670: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:38 vm10 ceph-mon[51861]: pgmap v670: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:40.126 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:40.126 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:40.155 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:40.155 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:40 vm08 ceph-mon[47734]: pgmap v671: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:40 vm10 ceph-mon[51861]: pgmap v671: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:42 vm08 ceph-mon[47734]: pgmap v672: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:42 vm10 ceph-mon[51861]: pgmap v672: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:15:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:15:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:15:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:15:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:44 vm08 ceph-mon[47734]: pgmap v673: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:45.156 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:45.157 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:44 vm10 ceph-mon[51861]: pgmap v673: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:45.182 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:45.182 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:47.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:46 vm08 ceph-mon[47734]: pgmap v674: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:46 vm10 ceph-mon[51861]: pgmap v674: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:48.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:47 vm08 ceph-mon[47734]: pgmap v675: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:48.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:47 vm10 ceph-mon[51861]: pgmap v675: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:50.184 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:50.184 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:50.210 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:50.210 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:50.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:50 vm08 ceph-mon[47734]: pgmap v676: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:50.481 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:50 vm10 ceph-mon[51861]: pgmap v676: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:15:51.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:15:51.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:15:51.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:15:51.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:15:51.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:15:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:15:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:15:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:15:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:15:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:15:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:15:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:52 vm08 ceph-mon[47734]: pgmap v677: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:52 vm10 ceph-mon[51861]: pgmap v677: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:54 vm08 ceph-mon[47734]: pgmap v678: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:54 vm10 ceph-mon[51861]: pgmap v678: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:55.212 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:15:55.212 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:15:55.243 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:15:55.243 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:15:57.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:56 vm08 ceph-mon[47734]: pgmap v679: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:56 vm10 ceph-mon[51861]: pgmap v679: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:15:58.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:15:57 vm08 ceph-mon[47734]: pgmap v680: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:15:58.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:15:57 vm10 ceph-mon[51861]: pgmap v680: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:00.245 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:00.246 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:00.273 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:00.273 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:00 vm08 ceph-mon[47734]: pgmap v681: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:00 vm10 ceph-mon[51861]: pgmap v681: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:02 vm08 ceph-mon[47734]: pgmap v682: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:02 vm10 ceph-mon[51861]: pgmap v682: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:04.960 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:04 vm10 ceph-mon[51861]: pgmap v683: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:04 vm08 ceph-mon[47734]: pgmap v683: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:05.275 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:05.276 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:05.302 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:05.302 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:06 vm08 ceph-mon[47734]: pgmap v684: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:06 vm10 ceph-mon[51861]: pgmap v684: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:08 vm08 ceph-mon[47734]: pgmap v685: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:09.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:08 vm10 ceph-mon[51861]: pgmap v685: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:10.304 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:10.304 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:10.329 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:10.330 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:10 vm08 ceph-mon[47734]: pgmap v686: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:10 vm10 ceph-mon[51861]: pgmap v686: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:12 vm08 ceph-mon[47734]: pgmap v687: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:12 vm10 ceph-mon[51861]: pgmap v687: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:14 vm08 ceph-mon[47734]: pgmap v688: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:14 vm10 ceph-mon[51861]: pgmap v688: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:15.331 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:15.332 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:15.359 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:15.359 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:17.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:16 vm08 ceph-mon[47734]: pgmap v689: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:16 vm10 ceph-mon[51861]: pgmap v689: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:18.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:17 vm08 ceph-mon[47734]: pgmap v690: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:18.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:17 vm10 ceph-mon[51861]: pgmap v690: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:20.360 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:20.361 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:20.386 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:20.386 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:20 vm08 ceph-mon[47734]: pgmap v691: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:20 vm10 ceph-mon[51861]: pgmap v691: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:22 vm08 ceph-mon[47734]: pgmap v692: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:22 vm10 ceph-mon[51861]: pgmap v692: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:24 vm08 ceph-mon[47734]: pgmap v693: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:24 vm10 ceph-mon[51861]: pgmap v693: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:25.388 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:25.388 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:25.413 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:25.414 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:26.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:26 vm08 ceph-mon[47734]: pgmap v694: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:26 vm10 ceph-mon[51861]: pgmap v694: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:28 vm08 ceph-mon[47734]: pgmap v695: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:28 vm10 ceph-mon[51861]: pgmap v695: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:30.415 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:30.416 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:30.461 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:30.462 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:30 vm10 ceph-mon[51861]: pgmap v696: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:31.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:30 vm08 ceph-mon[47734]: pgmap v696: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:31 vm10 ceph-mon[51861]: pgmap v697: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:32.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:31 vm08 ceph-mon[47734]: pgmap v697: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:34 vm10 ceph-mon[51861]: pgmap v698: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:34 vm08 ceph-mon[47734]: pgmap v698: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:35.463 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:35.464 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:35.489 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:35.490 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:36 vm10 ceph-mon[51861]: pgmap v699: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:36 vm08 ceph-mon[47734]: pgmap v699: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:38.734 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:38 vm08 ceph-mon[47734]: pgmap v700: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:38 vm10 ceph-mon[51861]: pgmap v700: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:40.491 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:40.492 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:40.536 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:40.536 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:40 vm08 ceph-mon[47734]: pgmap v701: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:40 vm10 ceph-mon[51861]: pgmap v701: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:42.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:41 vm08 ceph-mon[47734]: pgmap v702: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:42.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:41 vm10 ceph-mon[51861]: pgmap v702: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:16:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:16:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:16:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:16:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:44 vm08 ceph-mon[47734]: pgmap v703: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:44 vm10 ceph-mon[51861]: pgmap v703: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:45.538 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:45.538 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:45.632 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:45.632 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:47.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:46 vm08 ceph-mon[47734]: pgmap v704: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:46 vm10 ceph-mon[51861]: pgmap v704: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:48.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:47 vm08 ceph-mon[47734]: pgmap v705: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:48.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:47 vm10 ceph-mon[51861]: pgmap v705: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:50.634 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:50.635 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:50.660 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:50.660 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:50.759 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:50 vm08 ceph-mon[47734]: pgmap v706: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:50.760 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:50 vm10 ceph-mon[51861]: pgmap v706: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:51.379 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:16:51.379 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:16:51.379 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:51 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:16:51.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:16:51.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:16:51.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:51 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:16:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:52 vm08 ceph-mon[47734]: pgmap v707: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:16:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:16:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:16:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:16:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:16:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:52 vm10 ceph-mon[51861]: pgmap v707: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:16:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:16:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:16:53.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:16:53.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:16:54.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:53 vm08 ceph-mon[47734]: pgmap v708: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:54.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:53 vm10 ceph-mon[51861]: pgmap v708: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:55.661 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:16:55.662 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:16:55.687 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:16:55.688 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:16:57.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:56 vm08 ceph-mon[47734]: pgmap v709: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:56 vm10 ceph-mon[51861]: pgmap v709: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:16:58.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:16:57 vm10 ceph-mon[51861]: pgmap v710: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:16:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:16:57 vm08 ceph-mon[47734]: pgmap v710: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:00 vm08 ceph-mon[47734]: pgmap v711: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:00 vm10 ceph-mon[51861]: pgmap v711: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:00.689 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:00.690 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:00.717 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:00.717 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:02 vm08 ceph-mon[47734]: pgmap v712: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:02 vm10 ceph-mon[51861]: pgmap v712: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:03 vm08 ceph-mon[47734]: pgmap v713: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:04.060 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:03 vm10 ceph-mon[51861]: pgmap v713: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:05.719 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:05.719 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:05.767 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:05.767 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:06 vm08 ceph-mon[47734]: pgmap v714: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:06 vm10 ceph-mon[51861]: pgmap v714: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:08 vm08 ceph-mon[47734]: pgmap v715: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:08 vm10 ceph-mon[51861]: pgmap v715: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:10 vm08 ceph-mon[47734]: pgmap v716: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:10 vm10 ceph-mon[51861]: pgmap v716: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:10.769 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:10.769 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:10.797 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:10.798 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:12.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:11 vm08 ceph-mon[47734]: pgmap v717: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:11 vm10 ceph-mon[51861]: pgmap v717: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:14.473 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:14 vm10 ceph-mon[51861]: pgmap v718: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:14 vm08 ceph-mon[47734]: pgmap v718: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:15.800 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:15.800 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:15.832 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:15.832 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:16.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:16 vm08 ceph-mon[47734]: pgmap v719: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:16.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:16 vm10 ceph-mon[51861]: pgmap v719: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:18 vm08 ceph-mon[47734]: pgmap v720: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:18 vm10 ceph-mon[51861]: pgmap v720: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:20 vm08 ceph-mon[47734]: pgmap v721: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:20 vm10 ceph-mon[51861]: pgmap v721: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:20.834 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:20.834 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:20.863 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:20.863 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:22.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:21 vm08 ceph-mon[47734]: pgmap v722: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:22.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:21 vm10 ceph-mon[51861]: pgmap v722: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:24 vm08 ceph-mon[47734]: pgmap v723: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:24 vm10 ceph-mon[51861]: pgmap v723: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:25.865 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:25.865 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:25.890 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:25.891 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:26.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:26 vm08 ceph-mon[47734]: pgmap v724: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:26 vm10 ceph-mon[51861]: pgmap v724: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:28 vm08 ceph-mon[47734]: pgmap v725: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:28 vm10 ceph-mon[51861]: pgmap v725: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:30 vm08 ceph-mon[47734]: pgmap v726: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:30 vm10 ceph-mon[51861]: pgmap v726: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:30.893 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:30.893 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:30.920 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:30.920 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:32.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:31 vm08 ceph-mon[47734]: pgmap v727: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:31 vm10 ceph-mon[51861]: pgmap v727: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:34 vm08 ceph-mon[47734]: pgmap v728: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:34 vm10 ceph-mon[51861]: pgmap v728: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:35.922 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:35.922 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:35.948 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:35.948 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:36.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:36 vm08 ceph-mon[47734]: pgmap v729: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:36 vm10 ceph-mon[51861]: pgmap v729: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:38 vm08 ceph-mon[47734]: pgmap v730: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:38 vm10 ceph-mon[51861]: pgmap v730: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:40 vm08 ceph-mon[47734]: pgmap v731: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:40 vm10 ceph-mon[51861]: pgmap v731: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:40.950 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:40.950 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:40.976 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:40.977 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:42.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:41 vm08 ceph-mon[47734]: pgmap v732: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:42.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:41 vm10 ceph-mon[51861]: pgmap v732: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:17:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:17:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:17:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:17:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:44 vm08 ceph-mon[47734]: pgmap v733: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:44 vm10 ceph-mon[51861]: pgmap v733: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:45.978 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:45.979 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:46.005 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:46.006 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:46.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:46 vm08 ceph-mon[47734]: pgmap v734: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:46 vm10 ceph-mon[51861]: pgmap v734: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:48 vm10 ceph-mon[51861]: pgmap v735: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:48.741 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:48 vm08 ceph-mon[47734]: pgmap v735: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:50 vm10 ceph-mon[51861]: pgmap v736: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:50 vm08 ceph-mon[47734]: pgmap v736: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:51.007 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:51.008 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:51.034 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:51.034 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:52.258 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:51 vm08 ceph-mon[47734]: pgmap v737: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:52.258 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:51 vm10 ceph-mon[51861]: pgmap v737: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:17:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:17:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:17:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:17:53.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:17:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:17:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:17:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:17:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:17:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:17:54.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:53 vm08 ceph-mon[47734]: pgmap v738: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:54.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:53 vm10 ceph-mon[51861]: pgmap v738: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:56.036 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:17:56.036 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:17:56.064 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:17:56.065 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:17:56.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:56 vm08 ceph-mon[47734]: pgmap v739: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:56 vm10 ceph-mon[51861]: pgmap v739: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:17:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:17:58 vm08 ceph-mon[47734]: pgmap v740: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:17:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:17:58 vm10 ceph-mon[51861]: pgmap v740: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:00 vm08 ceph-mon[47734]: pgmap v741: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:00 vm10 ceph-mon[51861]: pgmap v741: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:01.066 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:01.067 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:01.092 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:01.093 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:02.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:01 vm08 ceph-mon[47734]: pgmap v742: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:01 vm10 ceph-mon[51861]: pgmap v742: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:04 vm08 ceph-mon[47734]: pgmap v743: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:04 vm10 ceph-mon[51861]: pgmap v743: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:06.095 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:06.095 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:06.121 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:06.122 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:06.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:06 vm08 ceph-mon[47734]: pgmap v744: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:06 vm10 ceph-mon[51861]: pgmap v744: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:08 vm08 ceph-mon[47734]: pgmap v745: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:08 vm10 ceph-mon[51861]: pgmap v745: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:10.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:10 vm08 ceph-mon[47734]: pgmap v746: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:10 vm10 ceph-mon[51861]: pgmap v746: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:11.123 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:11.124 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:11.190 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:11.190 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:12.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:11 vm08 ceph-mon[47734]: pgmap v747: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:11 vm10 ceph-mon[51861]: pgmap v747: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:14 vm08 ceph-mon[47734]: pgmap v748: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:14 vm10 ceph-mon[51861]: pgmap v748: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:16.192 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:16.193 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:16.240 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:16.241 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:16.769 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:16 vm08 ceph-mon[47734]: pgmap v749: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:16 vm10 ceph-mon[51861]: pgmap v749: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:18 vm08 ceph-mon[47734]: pgmap v750: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:18 vm10 ceph-mon[51861]: pgmap v750: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:20.558 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:20 vm10 ceph-mon[51861]: pgmap v751: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:20 vm08 ceph-mon[47734]: pgmap v751: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:21.242 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:21.243 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:21.269 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:21.269 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:22.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:21 vm08 ceph-mon[47734]: pgmap v752: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:22.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:21 vm10 ceph-mon[51861]: pgmap v752: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:24 vm08 ceph-mon[47734]: pgmap v753: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:24 vm10 ceph-mon[51861]: pgmap v753: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:26.271 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:26.271 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:26.307 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:26.308 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:26 vm10 ceph-mon[51861]: pgmap v754: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:26 vm08 ceph-mon[47734]: pgmap v754: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:28 vm10 ceph-mon[51861]: pgmap v755: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:28.745 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:28 vm08 ceph-mon[47734]: pgmap v755: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:30 vm10 ceph-mon[51861]: pgmap v756: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:30 vm08 ceph-mon[47734]: pgmap v756: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:31.309 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:31.309 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:31.339 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:31.340 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:32.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:31 vm08 ceph-mon[47734]: pgmap v757: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:31 vm10 ceph-mon[51861]: pgmap v757: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:34 vm08 ceph-mon[47734]: pgmap v758: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:34 vm10 ceph-mon[51861]: pgmap v758: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:36.341 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:36.341 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:36.368 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:36.368 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:36.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:36 vm08 ceph-mon[47734]: pgmap v759: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:36 vm10 ceph-mon[51861]: pgmap v759: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:38 vm08 ceph-mon[47734]: pgmap v760: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:38 vm10 ceph-mon[51861]: pgmap v760: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:40.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:40 vm08 ceph-mon[47734]: pgmap v761: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:40 vm10 ceph-mon[51861]: pgmap v761: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:41.369 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:41.370 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:41.397 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:41.397 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:42.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:41 vm08 ceph-mon[47734]: pgmap v762: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:42.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:41 vm10 ceph-mon[51861]: pgmap v762: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:18:43.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:43 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:18:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:18:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:18:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:44 vm08 ceph-mon[47734]: pgmap v763: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:44 vm10 ceph-mon[51861]: pgmap v763: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:46.399 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:46.399 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:46.425 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:46.426 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:46 vm08 ceph-mon[47734]: pgmap v764: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:46 vm10 ceph-mon[51861]: pgmap v764: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:48 vm08 ceph-mon[47734]: pgmap v765: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:48 vm10 ceph-mon[51861]: pgmap v765: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:50 vm08 ceph-mon[47734]: pgmap v766: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:50 vm10 ceph-mon[51861]: pgmap v766: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:51.427 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:51.428 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:51.452 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:51.452 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:52.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:51 vm08 ceph-mon[47734]: pgmap v767: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:52.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:51 vm10 ceph-mon[51861]: pgmap v767: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:52.797 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:18:52.797 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:18:52.797 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:18:52.797 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:18:52.797 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:18:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:18:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:18:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:18:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:18:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:52 vm08 ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:18:54.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:53 vm08 ceph-mon[47734]: pgmap v768: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:54.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:53 vm10 ceph-mon[51861]: pgmap v768: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:56.454 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:18:56.454 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:18:56.480 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:18:56.481 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:18:56.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:56 vm08 ceph-mon[47734]: pgmap v769: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:56 vm10 ceph-mon[51861]: pgmap v769: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:18:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:18:58 vm08 ceph-mon[47734]: pgmap v770: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:18:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:18:58 vm10 ceph-mon[51861]: pgmap v770: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:00 vm08 ceph-mon[47734]: pgmap v771: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:00 vm10 ceph-mon[51861]: pgmap v771: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:01.482 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:01.482 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:19:01.520 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:01.520 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:02.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:01 vm08 ceph-mon[47734]: pgmap v772: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:01 vm10 ceph-mon[51861]: pgmap v772: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:04 vm08 ceph-mon[47734]: pgmap v773: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:04 vm10 ceph-mon[51861]: pgmap v773: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:06.521 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:06.522 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:19:06.548 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:06.549 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:06.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:06 vm08 ceph-mon[47734]: pgmap v774: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:06 vm10 ceph-mon[51861]: pgmap v774: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:08 vm08 ceph-mon[47734]: pgmap v775: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:08 vm10 ceph-mon[51861]: pgmap v775: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:10 vm08 ceph-mon[47734]: pgmap v776: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:10 vm10 ceph-mon[51861]: pgmap v776: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:11.550 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:11.550 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:19:11.577 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:11.578 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:12.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:11 vm08 ceph-mon[47734]: pgmap v777: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:11 vm10 ceph-mon[51861]: pgmap v777: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:14 vm08 ceph-mon[47734]: pgmap v778: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:14 vm10 ceph-mon[51861]: pgmap v778: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:16.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:16 vm08 ceph-mon[47734]: pgmap v779: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:16.579 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:16.580 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:19:16.607 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:16.607 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:16 vm10 ceph-mon[51861]: pgmap v779: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:18 vm08 ceph-mon[47734]: pgmap v780: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:18 vm10 ceph-mon[51861]: pgmap v780: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:20 vm08 ceph-mon[47734]: pgmap v781: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:20 vm10 ceph-mon[51861]: pgmap v781: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:21.609 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:21.609 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:19:21.636 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:21.637 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:22.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:21 vm08 ceph-mon[47734]: pgmap v782: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:22.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:21 vm10 ceph-mon[51861]: pgmap v782: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:24 vm08 ceph-mon[47734]: pgmap v783: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:24 vm10 ceph-mon[51861]: pgmap v783: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:26.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:26 vm08 ceph-mon[47734]: pgmap v784: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:26.638 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:26.639 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:19:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:26 vm10 ceph-mon[51861]: pgmap v784: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:26.665 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:26.665 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:28 vm08 ceph-mon[47734]: pgmap v785: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:28 vm10 ceph-mon[51861]: pgmap v785: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:30 vm08 ceph-mon[47734]: pgmap v786: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:30 vm10 ceph-mon[51861]: pgmap v786: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:31.666 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:31.667 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08:/fake /mnt/foo -o sync 2026-03-09T14:19:31.692 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:31.693 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:32.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:31 vm08 ceph-mon[47734]: pgmap v787: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:31 vm10 ceph-mon[51861]: pgmap v787: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:34 vm08 ceph-mon[47734]: pgmap v788: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:34 vm10 ceph-mon[51861]: pgmap v788: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:36.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:36 vm08.local ceph-mon[47734]: pgmap v789: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:36 vm10 ceph-mon[51861]: pgmap v789: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:36.694 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:36.695 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:19:36.723 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:36.724 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:38 vm08.local ceph-mon[47734]: pgmap v790: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:38 vm10 ceph-mon[51861]: pgmap v790: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:40 vm08.local ceph-mon[47734]: pgmap v791: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:40 vm10 ceph-mon[51861]: pgmap v791: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:41.725 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:41.726 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:19:41.752 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:41.752 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:42.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:41 vm08.local ceph-mon[47734]: pgmap v792: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:42.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:41 vm10 ceph-mon[51861]: pgmap v792: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:19:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:19:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:19:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:43 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:19:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:44 vm08.local ceph-mon[47734]: pgmap v793: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:44 vm10 ceph-mon[51861]: pgmap v793: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:46 vm08.local ceph-mon[47734]: pgmap v794: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:46 vm10 ceph-mon[51861]: pgmap v794: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:46.754 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:46.754 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:19:46.781 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:46.782 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:48 vm08.local ceph-mon[47734]: pgmap v795: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:48 vm10 ceph-mon[51861]: pgmap v795: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:50 vm08.local ceph-mon[47734]: pgmap v796: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:50 vm10 ceph-mon[51861]: pgmap v796: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:51.783 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:51.783 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:19:51.810 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:51.811 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:52.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:51 vm10 ceph-mon[51861]: pgmap v797: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:52.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:51 vm08.local ceph-mon[47734]: pgmap v797: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:52 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:19:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:52 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:19:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:52 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:19:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:19:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:19:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:52 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:19:54.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:53 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:19:54.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:53 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:19:54.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:53 vm08.local ceph-mon[47734]: pgmap v798: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:54.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:19:54.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:53 vm10 ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:19:54.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:53 vm10 ceph-mon[51861]: pgmap v798: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:56.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:56 vm08.local ceph-mon[47734]: pgmap v799: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:56 vm10 ceph-mon[51861]: pgmap v799: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:19:56.812 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:19:56.812 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:19:56.837 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:19:56.838 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:19:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:19:58 vm08.local ceph-mon[47734]: pgmap v800: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:19:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:19:58 vm10 ceph-mon[51861]: pgmap v800: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:00 vm08.local ceph-mon[47734]: pgmap v801: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T14:20:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:00 vm10 ceph-mon[51861]: pgmap v801: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:00 vm10 ceph-mon[51861]: overall HEALTH_OK 2026-03-09T14:20:01.839 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:01.840 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:01.866 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:01.866 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:01 vm10 ceph-mon[51861]: pgmap v802: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:01 vm08.local ceph-mon[47734]: pgmap v802: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:04 vm08.local ceph-mon[47734]: pgmap v803: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:04 vm10 ceph-mon[51861]: pgmap v803: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:06.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:06 vm08.local ceph-mon[47734]: pgmap v804: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:06 vm10 ceph-mon[51861]: pgmap v804: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:06.868 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:06.868 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:06.895 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:06.896 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:08 vm08.local ceph-mon[47734]: pgmap v805: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:08 vm10 ceph-mon[51861]: pgmap v805: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:10 vm08.local ceph-mon[47734]: pgmap v806: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:10 vm10 ceph-mon[51861]: pgmap v806: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:11.898 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:11.898 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:11.925 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:11.925 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:11 vm10 ceph-mon[51861]: pgmap v807: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:11 vm08.local ceph-mon[47734]: pgmap v807: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:14 vm08.local ceph-mon[47734]: pgmap v808: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:14 vm10 ceph-mon[51861]: pgmap v808: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:16.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:16 vm08.local ceph-mon[47734]: pgmap v809: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:16 vm10 ceph-mon[51861]: pgmap v809: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:16.927 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:16.927 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:16.954 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:16.955 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:18 vm08.local ceph-mon[47734]: pgmap v810: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:18 vm10 ceph-mon[51861]: pgmap v810: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:20 vm08.local ceph-mon[47734]: pgmap v811: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:20 vm10 ceph-mon[51861]: pgmap v811: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:21.956 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:21.957 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:21.982 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:21.982 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:22.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:21 vm10 ceph-mon[51861]: pgmap v812: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:22.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:21 vm08.local ceph-mon[47734]: pgmap v812: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:24 vm08.local ceph-mon[47734]: pgmap v813: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:24.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:24 vm10.local ceph-mon[51861]: pgmap v813: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:26.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:26 vm08.local ceph-mon[47734]: pgmap v814: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:26 vm10.local ceph-mon[51861]: pgmap v814: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:26.984 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:26.985 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:27.011 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:27.011 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:28 vm08.local ceph-mon[47734]: pgmap v815: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:28 vm10.local ceph-mon[51861]: pgmap v815: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:30 vm08.local ceph-mon[47734]: pgmap v816: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:30 vm10.local ceph-mon[51861]: pgmap v816: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:32.013 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:32.013 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:32.038 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:32.038 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:31 vm10.local ceph-mon[51861]: pgmap v817: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:32.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:31 vm08.local ceph-mon[47734]: pgmap v817: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:34 vm08.local ceph-mon[47734]: pgmap v818: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:34.565 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:34 vm10.local ceph-mon[51861]: pgmap v818: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:36.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:36 vm08.local ceph-mon[47734]: pgmap v819: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:36 vm10.local ceph-mon[51861]: pgmap v819: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:37.040 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:37.040 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:37.067 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:37.068 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:38 vm08.local ceph-mon[47734]: pgmap v820: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:38 vm10.local ceph-mon[51861]: pgmap v820: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:40 vm08.local ceph-mon[47734]: pgmap v821: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:40 vm10.local ceph-mon[51861]: pgmap v821: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:42.070 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:42.070 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:42.096 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:42.096 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:42.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:41 vm10.local ceph-mon[51861]: pgmap v822: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:41 vm08.local ceph-mon[47734]: pgmap v822: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:20:43.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:20:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:20:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:20:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:44 vm10.local ceph-mon[51861]: pgmap v823: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:44 vm08.local ceph-mon[47734]: pgmap v823: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:46 vm08.local ceph-mon[47734]: pgmap v824: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:46 vm10.local ceph-mon[51861]: pgmap v824: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:47.097 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:47.098 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:47.123 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:47.124 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:48 vm08.local ceph-mon[47734]: pgmap v825: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:48 vm10.local ceph-mon[51861]: pgmap v825: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:50 vm08.local ceph-mon[47734]: pgmap v826: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:50 vm10.local ceph-mon[51861]: pgmap v826: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:52.125 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:52.126 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:52.151 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:52.152 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:52.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:51 vm10.local ceph-mon[51861]: pgmap v827: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:52.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:51 vm08.local ceph-mon[47734]: pgmap v827: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:53 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:20:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:53 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:20:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:53 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:20:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:53 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:20:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:53 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:20:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:53 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:20:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:54 vm08.local ceph-mon[47734]: pgmap v828: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:20:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:20:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:54 vm10.local ceph-mon[51861]: pgmap v828: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:20:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:20:56.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:56 vm08.local ceph-mon[47734]: pgmap v829: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:56 vm10.local ceph-mon[51861]: pgmap v829: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:20:57.153 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:20:57.153 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:20:57.182 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:20:57.183 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:20:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:20:58 vm08.local ceph-mon[47734]: pgmap v830: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:20:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:20:58 vm10.local ceph-mon[51861]: pgmap v830: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:00 vm08.local ceph-mon[47734]: pgmap v831: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:00 vm10.local ceph-mon[51861]: pgmap v831: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:01 vm10.local ceph-mon[51861]: pgmap v832: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:02.184 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:02.185 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:02.210 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:02.211 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:01 vm08.local ceph-mon[47734]: pgmap v832: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:04 vm08.local ceph-mon[47734]: pgmap v833: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:04 vm10.local ceph-mon[51861]: pgmap v833: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:06.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:06 vm08.local ceph-mon[47734]: pgmap v834: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:06 vm10.local ceph-mon[51861]: pgmap v834: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:07.212 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:07.212 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:07.238 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:07.239 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:08 vm08.local ceph-mon[47734]: pgmap v835: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:08 vm10.local ceph-mon[51861]: pgmap v835: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:10 vm08.local ceph-mon[47734]: pgmap v836: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:10 vm10.local ceph-mon[51861]: pgmap v836: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:11 vm10.local ceph-mon[51861]: pgmap v837: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:12.240 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:12.241 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:12.267 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:12.267 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:11 vm08.local ceph-mon[47734]: pgmap v837: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:14 vm08.local ceph-mon[47734]: pgmap v838: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:14 vm10.local ceph-mon[51861]: pgmap v838: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:16.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:16 vm08.local ceph-mon[47734]: pgmap v839: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:16 vm10.local ceph-mon[51861]: pgmap v839: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:17.269 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:17.269 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:17.300 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:17.300 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:18 vm08.local ceph-mon[47734]: pgmap v840: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:18 vm10.local ceph-mon[51861]: pgmap v840: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:20 vm08.local ceph-mon[47734]: pgmap v841: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:20 vm10.local ceph-mon[51861]: pgmap v841: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:22.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:21 vm10.local ceph-mon[51861]: pgmap v842: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:22.301 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:22.302 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:21 vm08.local ceph-mon[47734]: pgmap v842: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:22.329 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:22.329 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:24 vm08.local ceph-mon[47734]: pgmap v843: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:24 vm10.local ceph-mon[51861]: pgmap v843: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:26.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:26 vm08.local ceph-mon[47734]: pgmap v844: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:26 vm10.local ceph-mon[51861]: pgmap v844: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:27.331 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:27.331 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:27.358 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:27.358 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:28 vm08.local ceph-mon[47734]: pgmap v845: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:28 vm10.local ceph-mon[51861]: pgmap v845: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:30 vm08.local ceph-mon[47734]: pgmap v846: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:30 vm10.local ceph-mon[51861]: pgmap v846: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:31 vm10.local ceph-mon[51861]: pgmap v847: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:32.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:31 vm08.local ceph-mon[47734]: pgmap v847: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:32.360 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:32.360 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:32.387 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:32.387 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:34 vm08.local ceph-mon[47734]: pgmap v848: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:34 vm10.local ceph-mon[51861]: pgmap v848: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:36.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:36 vm08.local ceph-mon[47734]: pgmap v849: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:36 vm10.local ceph-mon[51861]: pgmap v849: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:37.389 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:37.389 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:37.415 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:37.415 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:38 vm08.local ceph-mon[47734]: pgmap v850: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:38 vm10.local ceph-mon[51861]: pgmap v850: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:40 vm08.local ceph-mon[47734]: pgmap v851: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:40 vm10.local ceph-mon[51861]: pgmap v851: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:42.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:41 vm10.local ceph-mon[51861]: pgmap v852: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:41 vm08.local ceph-mon[47734]: pgmap v852: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:42.416 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:42.417 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:42.444 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:42.445 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:21:43.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:21:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:21:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:21:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:44 vm10.local ceph-mon[51861]: pgmap v853: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:44 vm08.local ceph-mon[47734]: pgmap v853: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:46 vm08.local ceph-mon[47734]: pgmap v854: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:46 vm10.local ceph-mon[51861]: pgmap v854: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:47.446 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:47.446 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:47.471 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:47.472 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:48 vm08.local ceph-mon[47734]: pgmap v855: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:48 vm10.local ceph-mon[51861]: pgmap v855: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:50 vm08.local ceph-mon[47734]: pgmap v856: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:50 vm10.local ceph-mon[51861]: pgmap v856: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:52.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:51 vm10.local ceph-mon[51861]: pgmap v857: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:52.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:51 vm08.local ceph-mon[47734]: pgmap v857: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:52.473 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:52.474 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:52.499 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:52.499 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:53.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:53 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:21:53.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:53 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:21:53.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:53 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:21:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:53 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:21:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:53 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:21:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:53 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:21:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:54 vm08.local ceph-mon[47734]: pgmap v858: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:21:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:21:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:54 vm10.local ceph-mon[51861]: pgmap v858: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:21:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:21:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:56 vm08.local ceph-mon[47734]: pgmap v859: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:56 vm10.local ceph-mon[51861]: pgmap v859: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:21:57.501 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:21:57.501 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:21:57.527 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:21:57.527 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:21:58.765 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:21:58 vm08.local ceph-mon[47734]: pgmap v860: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:21:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:21:58 vm10.local ceph-mon[51861]: pgmap v860: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:00.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:00 vm08.local ceph-mon[47734]: pgmap v861: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:00 vm10.local ceph-mon[51861]: pgmap v861: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:01 vm10.local ceph-mon[51861]: pgmap v862: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:01 vm08.local ceph-mon[47734]: pgmap v862: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:02.529 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:02.529 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:02.556 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:02.557 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:04 vm08.local ceph-mon[47734]: pgmap v863: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:04 vm10.local ceph-mon[51861]: pgmap v863: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:06.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:06 vm08.local ceph-mon[47734]: pgmap v864: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:06 vm10.local ceph-mon[51861]: pgmap v864: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:07.559 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:07.560 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:07.588 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:07.589 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:08 vm08.local ceph-mon[47734]: pgmap v865: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:08 vm10.local ceph-mon[51861]: pgmap v865: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:10 vm08.local ceph-mon[47734]: pgmap v866: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:10 vm10.local ceph-mon[51861]: pgmap v866: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:11 vm10.local ceph-mon[51861]: pgmap v867: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:11 vm08.local ceph-mon[47734]: pgmap v867: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:12.590 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:12.590 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:12.618 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:12.619 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:14 vm08.local ceph-mon[47734]: pgmap v868: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:14 vm10.local ceph-mon[51861]: pgmap v868: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:16.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:16 vm08.local ceph-mon[47734]: pgmap v869: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:16 vm10.local ceph-mon[51861]: pgmap v869: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:17.620 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:17.620 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:17.645 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:17.646 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:18 vm08.local ceph-mon[47734]: pgmap v870: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:18 vm10.local ceph-mon[51861]: pgmap v870: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:20 vm10.local ceph-mon[51861]: pgmap v871: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:20.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:20 vm08.local ceph-mon[47734]: pgmap v871: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:22.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:21 vm10.local ceph-mon[51861]: pgmap v872: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:22.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:21 vm08.local ceph-mon[47734]: pgmap v872: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:22.647 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:22.648 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:22.674 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:22.674 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:24 vm08.local ceph-mon[47734]: pgmap v873: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:24 vm10.local ceph-mon[51861]: pgmap v873: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:26.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:26 vm08.local ceph-mon[47734]: pgmap v874: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:26 vm10.local ceph-mon[51861]: pgmap v874: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:27.675 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:27.676 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:27.701 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:27.701 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:28 vm08.local ceph-mon[47734]: pgmap v875: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:28 vm10.local ceph-mon[51861]: pgmap v875: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:30 vm08.local ceph-mon[47734]: pgmap v876: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:30 vm10.local ceph-mon[51861]: pgmap v876: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:31 vm10.local ceph-mon[51861]: pgmap v877: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:32.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:31 vm08.local ceph-mon[47734]: pgmap v877: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:32.703 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:32.703 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:32.745 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:32.746 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:34 vm08.local ceph-mon[47734]: pgmap v878: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:34 vm10.local ceph-mon[51861]: pgmap v878: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:36.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:36 vm08.local ceph-mon[47734]: pgmap v879: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:36 vm10.local ceph-mon[51861]: pgmap v879: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:37.747 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:37.747 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:37.772 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:37.773 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:38 vm08.local ceph-mon[47734]: pgmap v880: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:38 vm10.local ceph-mon[51861]: pgmap v880: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:40 vm08.local ceph-mon[47734]: pgmap v881: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:40 vm10.local ceph-mon[51861]: pgmap v881: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:42.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:41 vm10.local ceph-mon[51861]: pgmap v882: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:41 vm08.local ceph-mon[47734]: pgmap v882: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:42.774 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:42.775 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:42.801 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:42.801 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:22:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:22:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:22:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:22:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:44 vm10.local ceph-mon[51861]: pgmap v883: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:44 vm08.local ceph-mon[47734]: pgmap v883: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:46 vm08.local ceph-mon[47734]: pgmap v884: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:46 vm10.local ceph-mon[51861]: pgmap v884: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:47.802 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:47.803 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:47.828 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:47.829 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:48 vm08.local ceph-mon[47734]: pgmap v885: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:48 vm10.local ceph-mon[51861]: pgmap v885: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:50 vm08.local ceph-mon[47734]: pgmap v886: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:50 vm10.local ceph-mon[51861]: pgmap v886: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:52.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:51 vm10.local ceph-mon[51861]: pgmap v887: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:52.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:51 vm08.local ceph-mon[47734]: pgmap v887: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:52.830 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:52.831 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:52.856 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:52.857 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:54 vm08.local ceph-mon[47734]: pgmap v888: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:22:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:22:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:22:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:22:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:22:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:54 vm10.local ceph-mon[51861]: pgmap v888: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:22:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:22:54.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:22:54.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:22:54.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:22:56.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:56 vm08.local ceph-mon[47734]: pgmap v889: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:56 vm10.local ceph-mon[51861]: pgmap v889: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:22:57.858 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:22:57.859 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:22:57.885 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:22:57.885 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:22:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:22:58 vm08.local ceph-mon[47734]: pgmap v890: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:22:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:22:58 vm10.local ceph-mon[51861]: pgmap v890: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:00 vm10.local ceph-mon[51861]: pgmap v891: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:00 vm08.local ceph-mon[47734]: pgmap v891: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:01 vm10.local ceph-mon[51861]: pgmap v892: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:01 vm08.local ceph-mon[47734]: pgmap v892: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:02.886 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:02.887 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:02.913 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:02.913 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:04 vm08.local ceph-mon[47734]: pgmap v893: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:04 vm10.local ceph-mon[51861]: pgmap v893: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:06.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:06 vm08.local ceph-mon[47734]: pgmap v894: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:06 vm10.local ceph-mon[51861]: pgmap v894: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:07.914 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:07.915 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:07.940 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:07.941 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:08 vm08.local ceph-mon[47734]: pgmap v895: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:08 vm10.local ceph-mon[51861]: pgmap v895: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:10 vm08.local ceph-mon[47734]: pgmap v896: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:10 vm10.local ceph-mon[51861]: pgmap v896: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:11 vm10.local ceph-mon[51861]: pgmap v897: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:11 vm08.local ceph-mon[47734]: pgmap v897: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:12.942 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:12.942 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:12.967 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:12.968 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:14 vm08.local ceph-mon[47734]: pgmap v898: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:14 vm10.local ceph-mon[51861]: pgmap v898: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:16.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:16 vm08.local ceph-mon[47734]: pgmap v899: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:16 vm10.local ceph-mon[51861]: pgmap v899: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:17.969 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:17.969 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:18.124 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:18.124 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:18 vm10.local ceph-mon[51861]: pgmap v900: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:18.771 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:18 vm08.local ceph-mon[47734]: pgmap v900: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:20 vm10.local ceph-mon[51861]: pgmap v901: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:20.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:20 vm08.local ceph-mon[47734]: pgmap v901: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:22.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:21 vm10.local ceph-mon[51861]: pgmap v902: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:22.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:21 vm08.local ceph-mon[47734]: pgmap v902: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:23.126 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:23.126 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:23.152 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:23.153 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:24 vm10.local ceph-mon[51861]: pgmap v903: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:24 vm08.local ceph-mon[47734]: pgmap v903: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:26 vm10.local ceph-mon[51861]: pgmap v904: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:26.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:26 vm08.local ceph-mon[47734]: pgmap v904: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:28.154 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:28.155 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:28.181 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:28.181 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:28 vm10.local ceph-mon[51861]: pgmap v905: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:28.771 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:28 vm08.local ceph-mon[47734]: pgmap v905: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:30 vm10.local ceph-mon[51861]: pgmap v906: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:30 vm08.local ceph-mon[47734]: pgmap v906: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:32.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:31 vm10.local ceph-mon[51861]: pgmap v907: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:32.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:31 vm08.local ceph-mon[47734]: pgmap v907: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:33.183 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:33.183 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:33.210 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:33.211 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:34 vm10.local ceph-mon[51861]: pgmap v908: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:34 vm08.local ceph-mon[47734]: pgmap v908: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:36 vm10.local ceph-mon[51861]: pgmap v909: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:36 vm08.local ceph-mon[47734]: pgmap v909: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:38.213 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:38.214 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:38.238 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:38.239 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:38 vm10.local ceph-mon[51861]: pgmap v910: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:38.771 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:38 vm08.local ceph-mon[47734]: pgmap v910: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:40 vm10.local ceph-mon[51861]: pgmap v911: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:40 vm08.local ceph-mon[47734]: pgmap v911: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:42.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:41 vm10.local ceph-mon[51861]: pgmap v912: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:41 vm08.local ceph-mon[47734]: pgmap v912: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:43.240 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:43.241 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:43.267 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:43.268 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:23:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:23:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:23:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:23:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:44 vm10.local ceph-mon[51861]: pgmap v913: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:44 vm08.local ceph-mon[47734]: pgmap v913: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:46.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:46 vm08.local ceph-mon[47734]: pgmap v914: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:46 vm10.local ceph-mon[51861]: pgmap v914: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:48.269 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:48.270 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:48.296 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:48.297 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:48 vm10.local ceph-mon[51861]: pgmap v915: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:48.772 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:48 vm08.local ceph-mon[47734]: pgmap v915: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:50 vm10.local ceph-mon[51861]: pgmap v916: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:50 vm08.local ceph-mon[47734]: pgmap v916: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:52.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:51 vm10.local ceph-mon[51861]: pgmap v917: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:52.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:51 vm08.local ceph-mon[47734]: pgmap v917: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:53.299 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:53.299 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:53.330 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:53.330 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:54 vm08.local ceph-mon[47734]: pgmap v918: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:23:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:23:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:23:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:23:54.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:23:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:54 vm10.local ceph-mon[51861]: pgmap v918: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:23:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:23:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:23:54.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:23:54.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:23:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:56 vm10.local ceph-mon[51861]: pgmap v919: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:56 vm08.local ceph-mon[47734]: pgmap v919: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:23:58.332 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:23:58.332 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:23:58.360 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:23:58.361 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:23:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:23:58 vm10.local ceph-mon[51861]: pgmap v920: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:23:58.774 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:23:58 vm08.local ceph-mon[47734]: pgmap v920: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:00 vm10.local ceph-mon[51861]: pgmap v921: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:00 vm08.local ceph-mon[47734]: pgmap v921: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:01 vm10.local ceph-mon[51861]: pgmap v922: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:01 vm08.local ceph-mon[47734]: pgmap v922: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:03.363 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:03.363 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:03.390 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:03.390 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:04 vm08.local ceph-mon[47734]: pgmap v923: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:04 vm10.local ceph-mon[51861]: pgmap v923: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:06 vm10.local ceph-mon[51861]: pgmap v924: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:06.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:06 vm08.local ceph-mon[47734]: pgmap v924: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:08.392 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:08.393 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:08.420 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:08.421 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:08 vm10.local ceph-mon[51861]: pgmap v925: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:08.775 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:08 vm08.local ceph-mon[47734]: pgmap v925: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:10 vm10.local ceph-mon[51861]: pgmap v926: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:10 vm08.local ceph-mon[47734]: pgmap v926: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:12.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:11 vm10.local ceph-mon[51861]: pgmap v927: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:12.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:11 vm08.local ceph-mon[47734]: pgmap v927: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:13.422 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:13.423 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:13.448 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:13.449 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:14 vm08.local ceph-mon[47734]: pgmap v928: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:14 vm10.local ceph-mon[51861]: pgmap v928: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:16 vm10.local ceph-mon[51861]: pgmap v929: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:16 vm08.local ceph-mon[47734]: pgmap v929: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:18.450 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:18.451 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:18.477 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:18.478 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:18 vm10.local ceph-mon[51861]: pgmap v930: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:18.776 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:18 vm08.local ceph-mon[47734]: pgmap v930: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:20.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:20 vm10.local ceph-mon[51861]: pgmap v931: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:20.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:20 vm08.local ceph-mon[47734]: pgmap v931: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:22.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:21 vm10.local ceph-mon[51861]: pgmap v932: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:22.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:21 vm08.local ceph-mon[47734]: pgmap v932: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:23.479 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:23.479 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:23.505 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:23.506 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:24 vm10.local ceph-mon[51861]: pgmap v933: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:24 vm08.local ceph-mon[47734]: pgmap v933: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:26 vm10.local ceph-mon[51861]: pgmap v934: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:26 vm08.local ceph-mon[47734]: pgmap v934: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:28.508 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:28.508 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:28.534 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:28.534 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:28 vm10.local ceph-mon[51861]: pgmap v935: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:28.777 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:28 vm08.local ceph-mon[47734]: pgmap v935: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:30 vm10.local ceph-mon[51861]: pgmap v936: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:30 vm08.local ceph-mon[47734]: pgmap v936: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:32.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:32 vm08.local ceph-mon[47734]: pgmap v937: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:32.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:32 vm10.local ceph-mon[51861]: pgmap v937: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:33.536 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:33.536 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:33.562 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:33.562 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:34 vm10.local ceph-mon[51861]: pgmap v938: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:34 vm08.local ceph-mon[47734]: pgmap v938: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:36.609 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:36 vm08.local ceph-mon[47734]: pgmap v939: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:36 vm10.local ceph-mon[51861]: pgmap v939: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:38.564 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:38.564 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:38.589 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:38.589 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:38 vm10.local ceph-mon[51861]: pgmap v940: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:38.778 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:38 vm08.local ceph-mon[47734]: pgmap v940: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:40 vm10.local ceph-mon[51861]: pgmap v941: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:40 vm08.local ceph-mon[47734]: pgmap v941: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:42.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:41 vm08.local ceph-mon[47734]: pgmap v942: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:42.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:41 vm10.local ceph-mon[51861]: pgmap v942: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:24:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:24:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:24:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:24:43.590 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:43.591 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:43.616 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:43.616 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:44 vm10.local ceph-mon[51861]: pgmap v943: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:44 vm08.local ceph-mon[47734]: pgmap v943: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:46 vm10.local ceph-mon[51861]: pgmap v944: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:46 vm08.local ceph-mon[47734]: pgmap v944: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:48.618 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:48.618 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:48.643 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:48.644 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:48 vm10.local ceph-mon[51861]: pgmap v945: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:48.778 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:48 vm08.local ceph-mon[47734]: pgmap v945: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:50 vm10.local ceph-mon[51861]: pgmap v946: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:50 vm08.local ceph-mon[47734]: pgmap v946: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:52.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:51 vm08.local ceph-mon[47734]: pgmap v947: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:52.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:51 vm10.local ceph-mon[51861]: pgmap v947: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:53.645 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:53.646 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:53.671 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:53.672 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:54.315 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:54 vm10.local ceph-mon[51861]: pgmap v948: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:54.315 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:24:54.315 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:24:54.315 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:24:54.315 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:54.315 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:54 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:54 vm08.local ceph-mon[47734]: pgmap v948: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:24:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:24:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:24:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:54.619 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:54 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:55 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:55 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:55 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:55 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:56.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:55 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:56.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:55 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:56.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:55 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:56.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:55 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:24:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:56 vm08.local ceph-mon[47734]: pgmap v949: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:56 vm10.local ceph-mon[51861]: pgmap v949: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:24:58.673 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:24:58.673 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:24:58.698 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:24:58.699 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:24:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:24:58 vm10.local ceph-mon[51861]: pgmap v950: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:24:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:24:58 vm08.local ceph-mon[47734]: pgmap v950: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:00 vm10.local ceph-mon[51861]: pgmap v951: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:00 vm08.local ceph-mon[47734]: pgmap v951: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:02 vm08.local ceph-mon[47734]: pgmap v952: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:02 vm10.local ceph-mon[51861]: pgmap v952: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:03.700 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:03.701 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:03.725 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:03.726 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:04 vm10.local ceph-mon[51861]: pgmap v953: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:04 vm08.local ceph-mon[47734]: pgmap v953: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:06.609 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:06 vm08.local ceph-mon[47734]: pgmap v954: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:06 vm10.local ceph-mon[51861]: pgmap v954: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:08 vm10.local ceph-mon[51861]: pgmap v955: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:08.727 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:08.727 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:08.755 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:08.755 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:08.780 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:08 vm08.local ceph-mon[47734]: pgmap v955: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:10 vm10.local ceph-mon[51861]: pgmap v956: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:10 vm08.local ceph-mon[47734]: pgmap v956: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:12 vm10.local ceph-mon[51861]: pgmap v957: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:25:12.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:12 vm08.local ceph-mon[47734]: pgmap v957: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:25:13.756 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:13.757 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:13.795 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:13.796 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:14 vm10.local ceph-mon[51861]: pgmap v958: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:25:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:14 vm08.local ceph-mon[47734]: pgmap v958: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:25:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:16 vm10.local ceph-mon[51861]: pgmap v959: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:16 vm08.local ceph-mon[47734]: pgmap v959: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:18.780 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:18 vm08.local ceph-mon[47734]: pgmap v960: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:25:18.797 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:18.798 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:18.823 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:18.824 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:18 vm10.local ceph-mon[51861]: pgmap v960: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:25:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:20 vm08.local ceph-mon[47734]: pgmap v961: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:20 vm10.local ceph-mon[51861]: pgmap v961: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:22 vm08.local ceph-mon[47734]: pgmap v962: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:22 vm10.local ceph-mon[51861]: pgmap v962: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:23.825 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:23.826 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:23.851 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:23.852 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:24 vm10.local ceph-mon[51861]: pgmap v963: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:24 vm08.local ceph-mon[47734]: pgmap v963: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:26 vm10.local ceph-mon[51861]: pgmap v964: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:26 vm08.local ceph-mon[47734]: pgmap v964: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:28 vm10.local ceph-mon[51861]: pgmap v965: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:28.781 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:28 vm08.local ceph-mon[47734]: pgmap v965: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:28.853 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:28.854 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:28.897 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:28.897 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:30 vm10.local ceph-mon[51861]: pgmap v966: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:30 vm08.local ceph-mon[47734]: pgmap v966: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:32.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:32 vm10.local ceph-mon[51861]: pgmap v967: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:32 vm08.local ceph-mon[47734]: pgmap v967: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:33.899 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:33.899 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:33.944 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:33.944 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:34 vm10.local ceph-mon[51861]: pgmap v968: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:34 vm08.local ceph-mon[47734]: pgmap v968: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:36.609 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:36 vm08.local ceph-mon[47734]: pgmap v969: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:36 vm10.local ceph-mon[51861]: pgmap v969: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:38 vm10.local ceph-mon[51861]: pgmap v970: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:38.781 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:38 vm08.local ceph-mon[47734]: pgmap v970: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:38.946 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:38.946 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:38.972 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:38.973 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:40 vm08.local ceph-mon[47734]: pgmap v971: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:40 vm10.local ceph-mon[51861]: pgmap v971: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:42 vm08.local ceph-mon[47734]: pgmap v972: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:42 vm10.local ceph-mon[51861]: pgmap v972: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:25:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:25:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:25:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:25:43.974 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:43.975 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:44.002 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:44.003 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:44 vm08.local ceph-mon[47734]: pgmap v973: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:44 vm10.local ceph-mon[51861]: pgmap v973: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:46 vm10.local ceph-mon[51861]: pgmap v974: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:46 vm08.local ceph-mon[47734]: pgmap v974: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:48 vm10.local ceph-mon[51861]: pgmap v975: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:48.782 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:48 vm08.local ceph-mon[47734]: pgmap v975: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:49.004 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:49.004 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:49.029 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:49.030 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:50 vm10.local ceph-mon[51861]: pgmap v976: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:50 vm08.local ceph-mon[47734]: pgmap v976: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:52 vm08.local ceph-mon[47734]: pgmap v977: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:52 vm10.local ceph-mon[51861]: pgmap v977: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:54.031 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:54.032 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:54.059 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:54.060 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:25:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:54 vm10.local ceph-mon[51861]: pgmap v978: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:54.661 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:54 vm08.local ceph-mon[47734]: pgmap v978: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:55 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:25:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:55 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:25:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:55 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:25:55.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:55 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:25:55.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:55 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:25:55.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:55 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:25:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:56 vm10.local ceph-mon[51861]: pgmap v979: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:56.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:56 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:25:56.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:56 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:25:56.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:56 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:25:56.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:56 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:25:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:56 vm08.local ceph-mon[47734]: pgmap v979: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:25:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:56 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:25:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:56 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:25:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:56 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:25:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:56 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:25:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:25:58 vm10.local ceph-mon[51861]: pgmap v980: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:58.785 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:25:58 vm08.local ceph-mon[47734]: pgmap v980: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:25:59.061 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:25:59.062 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:25:59.092 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:25:59.092 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:00 vm10.local ceph-mon[51861]: pgmap v981: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:00 vm08.local ceph-mon[47734]: pgmap v981: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:02.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:02 vm08.local ceph-mon[47734]: pgmap v982: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:02 vm10.local ceph-mon[51861]: pgmap v982: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:04.094 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:04.094 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:04.120 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:04.121 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:04 vm10.local ceph-mon[51861]: pgmap v983: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:04 vm08.local ceph-mon[47734]: pgmap v983: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:06.609 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:06 vm08.local ceph-mon[47734]: pgmap v984: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:06 vm10.local ceph-mon[51861]: pgmap v984: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:08 vm10.local ceph-mon[51861]: pgmap v985: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:08.785 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:08 vm08.local ceph-mon[47734]: pgmap v985: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:09.122 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:09.123 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:09.149 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:09.150 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:10 vm10.local ceph-mon[51861]: pgmap v986: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:10 vm08.local ceph-mon[47734]: pgmap v986: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:12 vm08.local ceph-mon[47734]: pgmap v987: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:12 vm10.local ceph-mon[51861]: pgmap v987: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:14.151 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:14.152 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:14.177 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:14.178 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:14 vm10.local ceph-mon[51861]: pgmap v988: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:14 vm08.local ceph-mon[47734]: pgmap v988: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:16.634 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:16 vm08.local ceph-mon[47734]: pgmap v989: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:16 vm10.local ceph-mon[51861]: pgmap v989: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:18 vm10.local ceph-mon[51861]: pgmap v990: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:18.786 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:18 vm08.local ceph-mon[47734]: pgmap v990: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:19.179 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:19.180 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:19.205 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:19.206 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:20.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:20 vm10.local ceph-mon[51861]: pgmap v991: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:20 vm08.local ceph-mon[47734]: pgmap v991: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:22 vm08.local ceph-mon[47734]: pgmap v992: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:22 vm10.local ceph-mon[51861]: pgmap v992: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:24.207 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:24.207 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:24.233 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:24.234 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:24 vm10.local ceph-mon[51861]: pgmap v993: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:24 vm08.local ceph-mon[47734]: pgmap v993: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:26 vm10.local ceph-mon[51861]: pgmap v994: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:26 vm08.local ceph-mon[47734]: pgmap v994: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:28.787 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:28 vm08.local ceph-mon[47734]: pgmap v995: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:28 vm10.local ceph-mon[51861]: pgmap v995: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:29.235 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:29.236 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:29.261 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:29.262 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:30 vm08.local ceph-mon[47734]: pgmap v996: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:30 vm10.local ceph-mon[51861]: pgmap v996: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:32 vm08.local ceph-mon[47734]: pgmap v997: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:32 vm10.local ceph-mon[51861]: pgmap v997: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:34.263 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:34.264 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:34.289 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:34.289 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:34 vm10.local ceph-mon[51861]: pgmap v998: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:34 vm08.local ceph-mon[47734]: pgmap v998: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:36.609 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:36 vm08.local ceph-mon[47734]: pgmap v999: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:36 vm10.local ceph-mon[51861]: pgmap v999: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:38 vm10.local ceph-mon[51861]: pgmap v1000: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:38.787 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:38 vm08.local ceph-mon[47734]: pgmap v1000: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:39.291 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:39.291 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:39.316 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:39.316 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:40 vm10.local ceph-mon[51861]: pgmap v1001: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:40 vm08.local ceph-mon[47734]: pgmap v1001: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:42 vm08.local ceph-mon[47734]: pgmap v1002: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:42 vm10.local ceph-mon[51861]: pgmap v1002: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:26:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:26:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:26:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:26:44.318 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:44.318 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:44.348 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:44.348 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:44 vm10.local ceph-mon[51861]: pgmap v1003: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:44 vm08.local ceph-mon[47734]: pgmap v1003: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:46 vm10.local ceph-mon[51861]: pgmap v1004: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:46 vm08.local ceph-mon[47734]: pgmap v1004: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:48 vm10.local ceph-mon[51861]: pgmap v1005: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:48.789 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:48 vm08.local ceph-mon[47734]: pgmap v1005: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:49.349 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:49.350 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:49.375 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:49.375 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:50 vm10.local ceph-mon[51861]: pgmap v1006: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:50 vm08.local ceph-mon[47734]: pgmap v1006: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:52 vm08.local ceph-mon[47734]: pgmap v1007: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:52 vm10.local ceph-mon[51861]: pgmap v1007: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:54.377 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:54.377 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:54.402 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:54.402 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:54 vm10.local ceph-mon[51861]: pgmap v1008: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:54 vm08.local ceph-mon[47734]: pgmap v1008: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:56.389 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:56 vm08.local ceph-mon[47734]: pgmap v1009: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:56.389 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:56 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:26:56.389 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:56 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:26:56.389 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:56 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:26:56.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:56 vm10.local ceph-mon[51861]: pgmap v1009: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:26:56.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:56 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:26:56.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:56 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:26:56.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:56 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:26:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:58 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:26:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:58 vm08.local ceph-mon[47734]: pgmap v1010: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:58 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:26:58.559 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:58 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:26:58.559 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:58 vm10.local ceph-mon[51861]: pgmap v1010: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:26:58.559 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:58 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:26:59.404 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:26:59.404 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:26:59.436 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:26:59.436 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:26:59.886 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:26:59.886 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:26:59.886 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:26:59.886 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:26:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:27:00.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:27:00.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:27:00.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:27:00.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:26:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:27:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:00 vm10.local ceph-mon[51861]: pgmap v1011: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:00 vm08.local ceph-mon[47734]: pgmap v1011: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:02 vm08.local ceph-mon[47734]: pgmap v1012: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:02 vm10.local ceph-mon[51861]: pgmap v1012: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:04.438 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:04.438 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:04.465 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:04.466 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:04.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:04 vm10.local ceph-mon[51861]: pgmap v1013: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:04 vm08.local ceph-mon[47734]: pgmap v1013: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:06.609 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:06 vm08.local ceph-mon[47734]: pgmap v1014: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:06 vm10.local ceph-mon[51861]: pgmap v1014: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:08 vm08.local ceph-mon[47734]: pgmap v1015: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:08 vm10.local ceph-mon[51861]: pgmap v1015: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:09.467 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:09.468 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:09.493 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:09.494 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:10 vm10.local ceph-mon[51861]: pgmap v1016: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:10 vm08.local ceph-mon[47734]: pgmap v1016: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:12 vm08.local ceph-mon[47734]: pgmap v1017: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:12 vm10.local ceph-mon[51861]: pgmap v1017: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:14.495 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:14.496 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:14.522 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:14.522 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:14 vm10.local ceph-mon[51861]: pgmap v1018: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:14 vm08.local ceph-mon[47734]: pgmap v1018: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:16 vm10.local ceph-mon[51861]: pgmap v1019: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:16 vm08.local ceph-mon[47734]: pgmap v1019: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:18 vm10.local ceph-mon[51861]: pgmap v1020: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:18.791 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:18 vm08.local ceph-mon[47734]: pgmap v1020: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:19.524 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:19.524 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:19.549 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:19.549 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:20 vm10.local ceph-mon[51861]: pgmap v1021: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:20 vm08.local ceph-mon[47734]: pgmap v1021: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:22 vm08.local ceph-mon[47734]: pgmap v1022: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:22 vm10.local ceph-mon[51861]: pgmap v1022: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:24.551 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:24.551 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:24.578 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:24.579 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:24 vm10.local ceph-mon[51861]: pgmap v1023: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:24 vm08.local ceph-mon[47734]: pgmap v1023: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:26 vm10.local ceph-mon[51861]: pgmap v1024: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:26 vm08.local ceph-mon[47734]: pgmap v1024: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:28.792 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:28 vm08.local ceph-mon[47734]: pgmap v1025: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:28 vm10.local ceph-mon[51861]: pgmap v1025: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:29.580 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:29.581 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:29.607 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:29.608 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:30 vm08.local ceph-mon[47734]: pgmap v1026: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:30 vm10.local ceph-mon[51861]: pgmap v1026: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:32 vm08.local ceph-mon[47734]: pgmap v1027: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:32 vm10.local ceph-mon[51861]: pgmap v1027: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:34.609 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:34.609 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:34.635 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:34.635 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:34.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:34 vm10.local ceph-mon[51861]: pgmap v1028: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:34 vm08.local ceph-mon[47734]: pgmap v1028: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:36.608 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:36 vm08.local ceph-mon[47734]: pgmap v1029: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:36 vm10.local ceph-mon[51861]: pgmap v1029: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:38 vm10.local ceph-mon[51861]: pgmap v1030: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:38.793 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:38 vm08.local ceph-mon[47734]: pgmap v1030: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:39.637 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:39.637 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:39.663 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:39.664 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:40 vm10.local ceph-mon[51861]: pgmap v1031: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:40 vm08.local ceph-mon[47734]: pgmap v1031: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:42 vm08.local ceph-mon[47734]: pgmap v1032: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:42 vm10.local ceph-mon[51861]: pgmap v1032: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:27:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:27:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:27:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:27:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:44 vm08.local ceph-mon[47734]: pgmap v1033: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:44 vm10.local ceph-mon[51861]: pgmap v1033: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:44.665 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:44.666 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:44.692 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:44.693 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:46 vm10.local ceph-mon[51861]: pgmap v1034: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:46 vm08.local ceph-mon[47734]: pgmap v1034: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:48 vm10.local ceph-mon[51861]: pgmap v1035: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:48.793 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:48 vm08.local ceph-mon[47734]: pgmap v1035: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:49.694 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:49.694 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:49.719 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:49.720 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:50 vm10.local ceph-mon[51861]: pgmap v1036: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:50 vm08.local ceph-mon[47734]: pgmap v1036: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:52 vm08.local ceph-mon[47734]: pgmap v1037: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:52 vm10.local ceph-mon[51861]: pgmap v1037: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:54 vm10.local ceph-mon[51861]: pgmap v1038: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:54.721 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:54.721 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:54.746 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:54.747 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:54 vm08.local ceph-mon[47734]: pgmap v1038: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:56 vm10.local ceph-mon[51861]: pgmap v1039: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:56 vm08.local ceph-mon[47734]: pgmap v1039: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:27:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:58 vm10.local ceph-mon[51861]: pgmap v1040: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:58.669 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:58 vm08.local ceph-mon[47734]: pgmap v1040: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:27:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:27:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:27:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:27:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:27:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:27:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:27:59.748 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:27:59.748 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:27:59.775 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:27:59.775 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:27:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:27:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:27:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:27:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:27:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:27:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:28:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:00 vm10.local ceph-mon[51861]: pgmap v1041: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:00 vm08.local ceph-mon[47734]: pgmap v1041: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:02 vm08.local ceph-mon[47734]: pgmap v1042: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:02 vm10.local ceph-mon[51861]: pgmap v1042: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:04 vm10.local ceph-mon[51861]: pgmap v1043: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:04.776 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:04.777 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:04.802 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:04.802 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:04 vm08.local ceph-mon[47734]: pgmap v1043: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:06.608 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:06 vm08.local ceph-mon[47734]: pgmap v1044: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:06 vm10.local ceph-mon[51861]: pgmap v1044: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:08 vm10.local ceph-mon[51861]: pgmap v1045: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:08.795 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:08 vm08.local ceph-mon[47734]: pgmap v1045: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:09.803 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:09.804 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:09.829 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:09.830 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:10 vm10.local ceph-mon[51861]: pgmap v1046: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:10 vm08.local ceph-mon[47734]: pgmap v1046: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:12 vm08.local ceph-mon[47734]: pgmap v1047: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:12 vm10.local ceph-mon[51861]: pgmap v1047: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:14 vm10.local ceph-mon[51861]: pgmap v1048: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:14 vm08.local ceph-mon[47734]: pgmap v1048: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:14.831 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:14.832 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:14.857 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:14.858 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:16 vm10.local ceph-mon[51861]: pgmap v1049: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:16 vm08.local ceph-mon[47734]: pgmap v1049: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:18 vm10.local ceph-mon[51861]: pgmap v1050: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:18.796 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:18 vm08.local ceph-mon[47734]: pgmap v1050: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:19.859 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:19.859 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:19.885 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:19.885 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:20 vm10.local ceph-mon[51861]: pgmap v1051: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:20 vm08.local ceph-mon[47734]: pgmap v1051: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:22 vm08.local ceph-mon[47734]: pgmap v1052: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:22 vm10.local ceph-mon[51861]: pgmap v1052: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:24 vm10.local ceph-mon[51861]: pgmap v1053: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:24 vm08.local ceph-mon[47734]: pgmap v1053: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:24.886 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:24.887 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:24.912 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:24.913 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:26 vm10.local ceph-mon[51861]: pgmap v1054: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:26 vm08.local ceph-mon[47734]: pgmap v1054: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:28 vm10.local ceph-mon[51861]: pgmap v1055: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:28.797 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:28 vm08.local ceph-mon[47734]: pgmap v1055: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:29.914 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:29.915 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:29.940 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:29.940 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:30 vm10.local ceph-mon[51861]: pgmap v1056: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:30 vm08.local ceph-mon[47734]: pgmap v1056: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:32 vm08.local ceph-mon[47734]: pgmap v1057: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:32 vm10.local ceph-mon[51861]: pgmap v1057: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:34 vm10.local ceph-mon[51861]: pgmap v1058: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:34 vm08.local ceph-mon[47734]: pgmap v1058: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:34.942 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:34.942 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:34.966 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:34.967 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:36 vm10.local ceph-mon[51861]: pgmap v1059: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:36 vm08.local ceph-mon[47734]: pgmap v1059: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:38 vm10.local ceph-mon[51861]: pgmap v1060: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:38.797 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:38 vm08.local ceph-mon[47734]: pgmap v1060: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:39.968 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:39.969 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:39.994 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:39.995 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:40 vm10.local ceph-mon[51861]: pgmap v1061: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:40 vm08.local ceph-mon[47734]: pgmap v1061: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:42 vm08.local ceph-mon[47734]: pgmap v1062: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:42 vm10.local ceph-mon[51861]: pgmap v1062: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:28:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:28:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:28:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:28:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:44 vm08.local ceph-mon[47734]: pgmap v1063: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:44 vm10.local ceph-mon[51861]: pgmap v1063: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:44.996 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:44.997 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:45.022 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:45.023 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:46 vm10.local ceph-mon[51861]: pgmap v1064: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:46 vm08.local ceph-mon[47734]: pgmap v1064: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:48 vm10.local ceph-mon[51861]: pgmap v1065: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:48.798 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:48 vm08.local ceph-mon[47734]: pgmap v1065: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:50.024 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:50.025 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:50.050 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:50.051 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:50 vm10.local ceph-mon[51861]: pgmap v1066: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:50 vm08.local ceph-mon[47734]: pgmap v1066: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:52 vm08.local ceph-mon[47734]: pgmap v1067: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:52 vm10.local ceph-mon[51861]: pgmap v1067: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:54 vm10.local ceph-mon[51861]: pgmap v1068: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:54 vm08.local ceph-mon[47734]: pgmap v1068: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:55.052 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:28:55.053 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:28:55.083 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:28:55.084 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:28:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:56 vm10.local ceph-mon[51861]: pgmap v1069: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:56 vm08.local ceph-mon[47734]: pgmap v1069: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:28:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:58 vm10.local ceph-mon[51861]: pgmap v1070: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:58.800 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:58 vm08.local ceph-mon[47734]: pgmap v1070: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:28:59.378 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:28:59.378 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:28:59.378 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:28:59.378 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:28:59.378 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:28:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:28:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:28:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:28:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:28:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:28:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:28:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:29:00.085 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:00.086 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:00.111 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:00.112 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:00 vm10.local ceph-mon[51861]: pgmap v1071: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:00 vm08.local ceph-mon[47734]: pgmap v1071: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:02 vm08.local ceph-mon[47734]: pgmap v1072: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:02 vm10.local ceph-mon[51861]: pgmap v1072: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:04 vm10.local ceph-mon[51861]: pgmap v1073: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:04 vm08.local ceph-mon[47734]: pgmap v1073: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:05.113 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:05.114 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:05.139 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:05.140 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:06 vm10.local ceph-mon[51861]: pgmap v1074: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:06 vm08.local ceph-mon[47734]: pgmap v1074: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:08 vm10.local ceph-mon[51861]: pgmap v1075: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:08.800 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:08 vm08.local ceph-mon[47734]: pgmap v1075: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:10.141 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:10.142 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:10.168 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:10.169 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:10 vm10.local ceph-mon[51861]: pgmap v1076: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:10 vm08.local ceph-mon[47734]: pgmap v1076: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:12 vm08.local ceph-mon[47734]: pgmap v1077: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:12 vm10.local ceph-mon[51861]: pgmap v1077: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:14 vm10.local ceph-mon[51861]: pgmap v1078: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:14 vm08.local ceph-mon[47734]: pgmap v1078: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:15.170 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:15.171 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:15.301 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:15.302 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:16 vm08.local ceph-mon[47734]: pgmap v1079: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:16 vm10.local ceph-mon[51861]: pgmap v1079: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:18.801 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:18 vm08.local ceph-mon[47734]: pgmap v1080: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:18 vm10.local ceph-mon[51861]: pgmap v1080: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:20.303 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:20.304 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:20.329 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:20.330 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:20 vm08.local ceph-mon[47734]: pgmap v1081: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:20 vm10.local ceph-mon[51861]: pgmap v1081: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:22 vm08.local ceph-mon[47734]: pgmap v1082: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:22 vm10.local ceph-mon[51861]: pgmap v1082: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:24 vm10.local ceph-mon[51861]: pgmap v1083: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:24 vm08.local ceph-mon[47734]: pgmap v1083: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:25.331 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:25.331 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:25.357 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:25.357 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:26 vm10.local ceph-mon[51861]: pgmap v1084: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:26 vm08.local ceph-mon[47734]: pgmap v1084: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:28 vm10.local ceph-mon[51861]: pgmap v1085: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:28.803 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:28 vm08.local ceph-mon[47734]: pgmap v1085: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:30.359 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:30.359 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:30.385 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:30.385 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:30 vm10.local ceph-mon[51861]: pgmap v1086: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:30 vm08.local ceph-mon[47734]: pgmap v1086: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:32 vm08.local ceph-mon[47734]: pgmap v1087: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:32 vm10.local ceph-mon[51861]: pgmap v1087: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:34 vm10.local ceph-mon[51861]: pgmap v1088: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:34 vm08.local ceph-mon[47734]: pgmap v1088: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:35.386 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:35.387 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:35.414 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:35.415 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:36 vm10.local ceph-mon[51861]: pgmap v1089: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:36 vm08.local ceph-mon[47734]: pgmap v1089: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:38 vm10.local ceph-mon[51861]: pgmap v1090: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:38.804 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:38 vm08.local ceph-mon[47734]: pgmap v1090: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:40.416 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:40.416 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:40.443 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:40.443 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:40 vm10.local ceph-mon[51861]: pgmap v1091: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:40 vm08.local ceph-mon[47734]: pgmap v1091: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:42 vm08.local ceph-mon[47734]: pgmap v1092: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:42 vm10.local ceph-mon[51861]: pgmap v1092: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:29:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:29:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:29:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:29:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:44 vm08.local ceph-mon[47734]: pgmap v1093: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:44 vm10.local ceph-mon[51861]: pgmap v1093: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:45.445 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:45.445 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:45.472 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:45.472 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:46 vm10.local ceph-mon[51861]: pgmap v1094: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:46 vm08.local ceph-mon[47734]: pgmap v1094: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:48 vm10.local ceph-mon[51861]: pgmap v1095: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:48.805 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:48 vm08.local ceph-mon[47734]: pgmap v1095: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:50.474 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:50.474 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:50.499 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:50.500 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:50 vm10.local ceph-mon[51861]: pgmap v1096: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:50 vm08.local ceph-mon[47734]: pgmap v1096: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:52 vm08.local ceph-mon[47734]: pgmap v1097: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:52 vm10.local ceph-mon[51861]: pgmap v1097: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:54 vm10.local ceph-mon[51861]: pgmap v1098: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:54 vm08.local ceph-mon[47734]: pgmap v1098: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:55.501 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:29:55.502 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:29:55.526 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:29:55.527 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:29:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:56 vm10.local ceph-mon[51861]: pgmap v1099: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:56 vm08.local ceph-mon[47734]: pgmap v1099: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:29:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:58 vm10.local ceph-mon[51861]: pgmap v1100: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:58 vm08.local ceph-mon[47734]: pgmap v1100: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:29:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:29:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:29:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:29:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:29:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:29:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:29:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:29:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:29:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:29:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:29:59.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:29:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:30:00.528 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:00.528 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:00.554 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:00.554 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:00 vm10.local ceph-mon[51861]: pgmap v1101: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T14:30:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:00 vm08.local ceph-mon[47734]: pgmap v1101: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T14:30:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:02 vm08.local ceph-mon[47734]: pgmap v1102: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:02 vm10.local ceph-mon[51861]: pgmap v1102: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:04 vm10.local ceph-mon[51861]: pgmap v1103: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:04 vm08.local ceph-mon[47734]: pgmap v1103: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:05.556 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:05.556 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:05.581 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:05.582 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:06 vm10.local ceph-mon[51861]: pgmap v1104: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:06 vm08.local ceph-mon[47734]: pgmap v1104: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:08 vm10.local ceph-mon[51861]: pgmap v1105: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:08.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:08 vm08.local ceph-mon[47734]: pgmap v1105: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:10.583 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:10.584 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:10.609 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:10.609 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:10 vm10.local ceph-mon[51861]: pgmap v1106: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:10 vm08.local ceph-mon[47734]: pgmap v1106: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:12 vm08.local ceph-mon[47734]: pgmap v1107: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:12 vm10.local ceph-mon[51861]: pgmap v1107: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:14 vm10.local ceph-mon[51861]: pgmap v1108: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:14 vm08.local ceph-mon[47734]: pgmap v1108: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:15.610 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:15.611 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:15.636 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:15.636 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:16 vm10.local ceph-mon[51861]: pgmap v1109: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:16 vm08.local ceph-mon[47734]: pgmap v1109: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:18 vm10.local ceph-mon[51861]: pgmap v1110: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:18.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:18 vm08.local ceph-mon[47734]: pgmap v1110: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:20.638 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:20.638 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:20 vm10.local ceph-mon[51861]: pgmap v1111: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:20.663 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:20.664 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:20.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:20 vm08.local ceph-mon[47734]: pgmap v1111: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:22 vm08.local ceph-mon[47734]: pgmap v1112: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:22 vm10.local ceph-mon[51861]: pgmap v1112: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:24 vm10.local ceph-mon[51861]: pgmap v1113: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:24 vm08.local ceph-mon[47734]: pgmap v1113: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:25.665 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:25.665 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:25.691 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:25.691 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:26 vm10.local ceph-mon[51861]: pgmap v1114: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:26 vm08.local ceph-mon[47734]: pgmap v1114: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:28 vm10.local ceph-mon[51861]: pgmap v1115: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:28.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:28 vm08.local ceph-mon[47734]: pgmap v1115: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:30 vm10.local ceph-mon[51861]: pgmap v1116: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:30.692 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:30.693 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:30.718 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:30.718 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:30 vm08.local ceph-mon[47734]: pgmap v1116: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:32 vm08.local ceph-mon[47734]: pgmap v1117: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:32 vm10.local ceph-mon[51861]: pgmap v1117: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:34 vm10.local ceph-mon[51861]: pgmap v1118: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:34 vm08.local ceph-mon[47734]: pgmap v1118: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:35.720 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:35.720 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:35.746 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:35.747 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:36 vm10.local ceph-mon[51861]: pgmap v1119: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:36 vm08.local ceph-mon[47734]: pgmap v1119: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:38 vm10.local ceph-mon[51861]: pgmap v1120: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:38 vm08.local ceph-mon[47734]: pgmap v1120: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:40 vm10.local ceph-mon[51861]: pgmap v1121: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:40.749 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:40.749 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:40.777 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:40.777 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:40.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:40 vm08.local ceph-mon[47734]: pgmap v1121: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:42 vm08.local ceph-mon[47734]: pgmap v1122: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:42 vm10.local ceph-mon[51861]: pgmap v1122: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:30:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:30:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:30:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:30:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:44 vm08.local ceph-mon[47734]: pgmap v1123: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:44 vm10.local ceph-mon[51861]: pgmap v1123: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:45.778 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:45.779 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:45.804 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:45.804 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:46 vm10.local ceph-mon[51861]: pgmap v1124: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:46 vm08.local ceph-mon[47734]: pgmap v1124: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:48 vm10.local ceph-mon[51861]: pgmap v1125: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:48.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:48 vm08.local ceph-mon[47734]: pgmap v1125: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:50 vm10.local ceph-mon[51861]: pgmap v1126: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:50.806 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:50.806 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:50 vm08.local ceph-mon[47734]: pgmap v1126: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:50.833 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:50.834 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:52 vm08.local ceph-mon[47734]: pgmap v1127: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:52 vm10.local ceph-mon[51861]: pgmap v1127: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:54 vm10.local ceph-mon[51861]: pgmap v1128: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:54 vm08.local ceph-mon[47734]: pgmap v1128: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:55.835 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:30:55.835 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:30:55.889 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:30:55.889 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:30:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:56 vm10.local ceph-mon[51861]: pgmap v1129: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:56 vm08.local ceph-mon[47734]: pgmap v1129: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:30:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:58 vm10.local ceph-mon[51861]: pgmap v1130: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:58 vm08.local ceph-mon[47734]: pgmap v1130: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:30:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:30:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:30:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:30:59 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:30:59.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:30:59.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:30:59.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:30:59 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:31:00.891 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:00.891 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:00 vm10.local ceph-mon[51861]: pgmap v1131: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:31:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:31:00.916 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:00.917 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:00 vm08.local ceph-mon[47734]: pgmap v1131: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:31:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:31:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:02 vm08.local ceph-mon[47734]: pgmap v1132: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:02 vm10.local ceph-mon[51861]: pgmap v1132: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:04 vm10.local ceph-mon[51861]: pgmap v1133: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:04.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:04 vm08.local ceph-mon[47734]: pgmap v1133: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:05.918 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:05.918 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:05.943 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:05.944 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:06 vm10.local ceph-mon[51861]: pgmap v1134: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:06 vm08.local ceph-mon[47734]: pgmap v1134: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:08 vm08.local ceph-mon[47734]: pgmap v1135: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:08 vm10.local ceph-mon[51861]: pgmap v1135: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:10 vm10.local ceph-mon[51861]: pgmap v1136: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:10 vm08.local ceph-mon[47734]: pgmap v1136: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:10.946 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:10.946 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:10.972 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:10.973 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:12 vm08.local ceph-mon[47734]: pgmap v1137: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:12 vm10.local ceph-mon[51861]: pgmap v1137: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:14 vm10.local ceph-mon[51861]: pgmap v1138: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:14 vm08.local ceph-mon[47734]: pgmap v1138: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:15.974 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:15.975 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:16.001 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:16.002 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:16 vm10.local ceph-mon[51861]: pgmap v1139: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:16 vm08.local ceph-mon[47734]: pgmap v1139: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:18 vm08.local ceph-mon[47734]: pgmap v1140: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:18 vm10.local ceph-mon[51861]: pgmap v1140: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:20 vm08.local ceph-mon[47734]: pgmap v1141: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:20 vm10.local ceph-mon[51861]: pgmap v1141: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:21.003 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:21.004 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:21.029 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:21.030 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:22 vm08.local ceph-mon[47734]: pgmap v1142: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:22 vm10.local ceph-mon[51861]: pgmap v1142: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:24 vm10.local ceph-mon[51861]: pgmap v1143: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:24 vm08.local ceph-mon[47734]: pgmap v1143: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:26.031 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:26.032 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:26.059 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:26.059 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:26 vm08.local ceph-mon[47734]: pgmap v1144: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:26 vm10.local ceph-mon[51861]: pgmap v1144: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:28.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:28 vm08.local ceph-mon[47734]: pgmap v1145: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:28 vm10.local ceph-mon[51861]: pgmap v1145: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:30 vm08.local ceph-mon[47734]: pgmap v1146: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:30 vm10.local ceph-mon[51861]: pgmap v1146: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:31.060 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:31.061 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:31.087 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:31.088 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:32 vm08.local ceph-mon[47734]: pgmap v1147: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:32 vm10.local ceph-mon[51861]: pgmap v1147: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:34 vm10.local ceph-mon[51861]: pgmap v1148: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:34.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:34 vm08.local ceph-mon[47734]: pgmap v1148: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:36.089 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:36.089 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:36.115 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:36.115 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:36 vm10.local ceph-mon[51861]: pgmap v1149: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:36 vm08.local ceph-mon[47734]: pgmap v1149: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:38 vm10.local ceph-mon[51861]: pgmap v1150: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:38 vm08.local ceph-mon[47734]: pgmap v1150: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:40 vm08.local ceph-mon[47734]: pgmap v1151: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:40 vm10.local ceph-mon[51861]: pgmap v1151: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:41.117 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:41.118 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:41.144 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:41.144 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:42 vm08.local ceph-mon[47734]: pgmap v1152: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:42 vm10.local ceph-mon[51861]: pgmap v1152: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:31:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:31:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:31:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:31:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:44 vm08.local ceph-mon[47734]: pgmap v1153: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:44 vm10.local ceph-mon[51861]: pgmap v1153: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:46.145 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:46.146 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:46.171 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:46.172 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:46 vm10.local ceph-mon[51861]: pgmap v1154: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:46 vm08.local ceph-mon[47734]: pgmap v1154: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:48 vm10.local ceph-mon[51861]: pgmap v1155: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:48 vm08.local ceph-mon[47734]: pgmap v1155: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:50 vm08.local ceph-mon[47734]: pgmap v1156: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:50 vm10.local ceph-mon[51861]: pgmap v1156: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:51.173 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:51.174 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:51.200 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:51.200 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:52.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:52 vm08.local ceph-mon[47734]: pgmap v1157: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:52 vm10.local ceph-mon[51861]: pgmap v1157: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:54 vm10.local ceph-mon[51861]: pgmap v1158: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:54 vm08.local ceph-mon[47734]: pgmap v1158: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:56.202 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:31:56.202 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:31:56.229 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:31:56.229 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:31:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:56 vm10.local ceph-mon[51861]: pgmap v1159: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:56 vm08.local ceph-mon[47734]: pgmap v1159: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:31:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:31:58 vm08.local ceph-mon[47734]: pgmap v1160: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:31:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:31:58 vm10.local ceph-mon[51861]: pgmap v1160: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:00 vm08.local ceph-mon[47734]: pgmap v1161: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:32:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:32:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:32:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:32:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:32:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:00 vm10.local ceph-mon[51861]: pgmap v1161: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:32:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:32:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:32:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:32:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:32:01.231 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:01.231 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:01.257 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:01.258 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:02 vm08.local ceph-mon[47734]: pgmap v1162: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:02 vm10.local ceph-mon[51861]: pgmap v1162: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:04 vm10.local ceph-mon[51861]: pgmap v1163: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:04.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:04 vm08.local ceph-mon[47734]: pgmap v1163: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:06.260 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:06.260 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:06.287 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:06.287 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:06 vm10.local ceph-mon[51861]: pgmap v1164: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:06 vm08.local ceph-mon[47734]: pgmap v1164: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:08.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:08 vm08.local ceph-mon[47734]: pgmap v1165: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:08 vm10.local ceph-mon[51861]: pgmap v1165: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:10 vm08.local ceph-mon[47734]: pgmap v1166: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:10 vm10.local ceph-mon[51861]: pgmap v1166: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:11.288 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:11.289 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:11.315 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:11.316 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:12 vm10.local ceph-mon[51861]: pgmap v1167: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:12.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:12 vm08.local ceph-mon[47734]: pgmap v1167: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:14 vm10.local ceph-mon[51861]: pgmap v1168: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:14 vm08.local ceph-mon[47734]: pgmap v1168: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:16.317 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:16.318 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:16.343 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:16.344 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:16 vm10.local ceph-mon[51861]: pgmap v1169: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:16 vm08.local ceph-mon[47734]: pgmap v1169: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:18 vm08.local ceph-mon[47734]: pgmap v1170: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:18 vm10.local ceph-mon[51861]: pgmap v1170: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:20 vm08.local ceph-mon[47734]: pgmap v1171: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:20 vm10.local ceph-mon[51861]: pgmap v1171: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:21.345 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:21.346 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:21.373 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:21.373 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:22 vm10.local ceph-mon[51861]: pgmap v1172: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:22.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:22 vm08.local ceph-mon[47734]: pgmap v1172: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:24 vm10.local ceph-mon[51861]: pgmap v1173: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:24.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:24 vm08.local ceph-mon[47734]: pgmap v1173: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:26.375 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:26.375 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:26.403 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:26.403 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:26 vm08.local ceph-mon[47734]: pgmap v1174: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:26 vm10.local ceph-mon[51861]: pgmap v1174: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:28.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:28 vm08.local ceph-mon[47734]: pgmap v1175: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:28 vm10.local ceph-mon[51861]: pgmap v1175: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:30.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:30 vm08.local ceph-mon[47734]: pgmap v1176: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:30 vm10.local ceph-mon[51861]: pgmap v1176: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:31.405 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:31.406 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:31.432 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:31.433 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:32 vm10.local ceph-mon[51861]: pgmap v1177: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:32 vm08.local ceph-mon[47734]: pgmap v1177: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:34.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:34 vm08.local ceph-mon[47734]: pgmap v1178: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:34 vm10.local ceph-mon[51861]: pgmap v1178: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:36.434 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:36.435 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:36.460 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:36.460 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:36 vm08.local ceph-mon[47734]: pgmap v1179: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:36 vm10.local ceph-mon[51861]: pgmap v1179: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:38 vm08.local ceph-mon[47734]: pgmap v1180: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:38 vm10.local ceph-mon[51861]: pgmap v1180: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:40 vm08.local ceph-mon[47734]: pgmap v1181: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:40 vm10.local ceph-mon[51861]: pgmap v1181: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:41.462 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:41.462 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:41.488 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:41.488 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:42 vm10.local ceph-mon[51861]: pgmap v1182: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:42.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:42 vm08.local ceph-mon[47734]: pgmap v1182: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:32:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:32:43.698 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:32:43.698 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:32:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:44 vm10.local ceph-mon[51861]: pgmap v1183: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:44 vm08.local ceph-mon[47734]: pgmap v1183: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:46.489 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:46.490 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:46.516 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:46.516 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:46 vm10.local ceph-mon[51861]: pgmap v1184: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:46 vm08.local ceph-mon[47734]: pgmap v1184: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:48 vm08.local ceph-mon[47734]: pgmap v1185: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:48 vm10.local ceph-mon[51861]: pgmap v1185: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:50 vm08.local ceph-mon[47734]: pgmap v1186: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:50 vm10.local ceph-mon[51861]: pgmap v1186: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:51.518 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:51.518 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:51.543 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:51.544 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:52.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:52 vm10.local ceph-mon[51861]: pgmap v1187: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:52 vm08.local ceph-mon[47734]: pgmap v1187: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:54.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:54 vm08.local ceph-mon[47734]: pgmap v1188: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:54 vm10.local ceph-mon[51861]: pgmap v1188: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:56.546 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:32:56.546 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:32:56.573 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:32:56.574 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:32:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:56 vm08.local ceph-mon[47734]: pgmap v1189: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:56 vm10.local ceph-mon[51861]: pgmap v1189: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:32:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:32:58 vm08.local ceph-mon[47734]: pgmap v1190: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:32:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:32:58 vm10.local ceph-mon[51861]: pgmap v1190: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:00 vm08.local ceph-mon[47734]: pgmap v1191: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:33:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:33:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:33:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:33:00.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:33:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:00 vm10.local ceph-mon[51861]: pgmap v1191: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:33:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:33:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:33:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:33:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:33:01.576 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:01.577 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:01.604 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:01.604 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:02 vm10.local ceph-mon[51861]: pgmap v1192: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:02.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:02 vm08.local ceph-mon[47734]: pgmap v1192: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:04.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:04 vm08.local ceph-mon[47734]: pgmap v1193: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:04.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:04 vm10.local ceph-mon[51861]: pgmap v1193: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:06.606 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:06.606 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:06.635 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:06.636 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:06.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:06 vm08.local ceph-mon[47734]: pgmap v1194: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:06 vm10.local ceph-mon[51861]: pgmap v1194: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:08.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:08 vm08.local ceph-mon[47734]: pgmap v1195: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:08 vm10.local ceph-mon[51861]: pgmap v1195: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:10.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:10 vm08.local ceph-mon[47734]: pgmap v1196: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:10 vm10.local ceph-mon[51861]: pgmap v1196: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:11.637 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:11.638 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:11.664 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:11.665 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:12 vm10.local ceph-mon[51861]: pgmap v1197: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:12.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:12 vm08.local ceph-mon[47734]: pgmap v1197: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:14 vm08.local ceph-mon[47734]: pgmap v1198: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:14 vm10.local ceph-mon[51861]: pgmap v1198: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:16.666 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:16.667 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:16.693 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:16.694 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:16 vm08.local ceph-mon[47734]: pgmap v1199: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:16 vm10.local ceph-mon[51861]: pgmap v1199: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:18 vm08.local ceph-mon[47734]: pgmap v1200: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:18 vm10.local ceph-mon[51861]: pgmap v1200: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:20 vm08.local ceph-mon[47734]: pgmap v1201: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:20 vm10.local ceph-mon[51861]: pgmap v1201: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:21.695 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:21.695 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:21.722 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:21.722 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:22 vm10.local ceph-mon[51861]: pgmap v1202: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:22.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:22 vm08.local ceph-mon[47734]: pgmap v1202: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:24 vm08.local ceph-mon[47734]: pgmap v1203: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:24 vm10.local ceph-mon[51861]: pgmap v1203: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:26.724 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:26.724 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:26.753 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:26.753 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:26.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:26 vm08.local ceph-mon[47734]: pgmap v1204: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:26 vm10.local ceph-mon[51861]: pgmap v1204: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:28.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:28 vm08.local ceph-mon[47734]: pgmap v1205: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:28 vm10.local ceph-mon[51861]: pgmap v1205: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:30.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:30 vm08.local ceph-mon[47734]: pgmap v1206: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:30 vm10.local ceph-mon[51861]: pgmap v1206: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:31.755 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:31.755 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:31.783 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:31.784 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:32 vm10.local ceph-mon[51861]: pgmap v1207: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:32 vm08.local ceph-mon[47734]: pgmap v1207: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:34.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:34 vm08.local ceph-mon[47734]: pgmap v1208: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:34.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:34 vm10.local ceph-mon[51861]: pgmap v1208: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:36.785 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:36.786 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:36 vm08.local ceph-mon[47734]: pgmap v1209: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:36.814 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:36.815 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:36 vm10.local ceph-mon[51861]: pgmap v1209: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:38 vm08.local ceph-mon[47734]: pgmap v1210: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:38 vm10.local ceph-mon[51861]: pgmap v1210: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:40.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:40 vm08.local ceph-mon[47734]: pgmap v1211: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:40.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:40 vm10.local ceph-mon[51861]: pgmap v1211: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:41.816 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:41.816 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:41.843 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:41.844 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:42 vm10.local ceph-mon[51861]: pgmap v1212: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:42.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:42 vm08.local ceph-mon[47734]: pgmap v1212: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:33:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:33:43.704 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:33:43.704 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:33:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:44 vm10.local ceph-mon[51861]: pgmap v1213: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:44.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:44 vm08.local ceph-mon[47734]: pgmap v1213: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:46 vm08.local ceph-mon[47734]: pgmap v1214: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:46.845 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:46.845 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:46.871 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:46.872 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:46 vm10.local ceph-mon[51861]: pgmap v1214: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:48 vm08.local ceph-mon[47734]: pgmap v1215: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:48 vm10.local ceph-mon[51861]: pgmap v1215: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:50.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:50 vm08.local ceph-mon[47734]: pgmap v1216: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:50 vm10.local ceph-mon[51861]: pgmap v1216: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:51.873 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:51.873 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:51.909 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:51.910 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:52 vm08.local ceph-mon[47734]: pgmap v1217: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:52 vm10.local ceph-mon[51861]: pgmap v1217: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:54 vm08.local ceph-mon[47734]: pgmap v1218: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:54 vm10.local ceph-mon[51861]: pgmap v1218: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:56 vm08.local ceph-mon[47734]: pgmap v1219: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:56 vm10.local ceph-mon[51861]: pgmap v1219: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:33:56.911 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:33:56.912 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:33:56.937 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:33:56.937 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:33:58.827 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:33:58 vm08.local ceph-mon[47734]: pgmap v1220: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:33:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:33:58 vm10.local ceph-mon[51861]: pgmap v1220: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:00 vm10.local ceph-mon[51861]: pgmap v1221: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:34:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:34:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:34:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:34:00.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:34:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:00 vm08.local ceph-mon[47734]: pgmap v1221: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:34:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:34:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:34:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:34:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:34:01.939 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:01.940 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:01.965 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:01.965 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:02 vm10.local ceph-mon[51861]: pgmap v1222: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:02.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:02 vm08.local ceph-mon[47734]: pgmap v1222: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:04 vm08.local ceph-mon[47734]: pgmap v1223: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:04 vm10.local ceph-mon[51861]: pgmap v1223: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:06 vm08.local ceph-mon[47734]: pgmap v1224: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:06 vm10.local ceph-mon[51861]: pgmap v1224: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:06.967 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:06.967 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:06.995 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:06.996 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:08.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:08 vm08.local ceph-mon[47734]: pgmap v1225: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:08 vm10.local ceph-mon[51861]: pgmap v1225: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:10.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:10 vm08.local ceph-mon[47734]: pgmap v1226: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:10 vm10.local ceph-mon[51861]: pgmap v1226: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:11.997 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:11.998 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:12.024 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:12.024 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:12 vm10.local ceph-mon[51861]: pgmap v1227: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:12.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:12 vm08.local ceph-mon[47734]: pgmap v1227: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:14 vm08.local ceph-mon[47734]: pgmap v1228: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:14 vm10.local ceph-mon[51861]: pgmap v1228: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:16 vm08.local ceph-mon[47734]: pgmap v1229: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:16 vm10.local ceph-mon[51861]: pgmap v1229: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:17.026 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:17.027 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:17.060 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:17.061 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:18 vm08.local ceph-mon[47734]: pgmap v1230: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:18 vm10.local ceph-mon[51861]: pgmap v1230: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:20.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:20 vm08.local ceph-mon[47734]: pgmap v1231: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:20 vm10.local ceph-mon[51861]: pgmap v1231: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:22.063 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:22.063 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:22.089 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:22.090 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:22 vm10.local ceph-mon[51861]: pgmap v1232: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:22.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:22 vm08.local ceph-mon[47734]: pgmap v1232: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:24 vm08.local ceph-mon[47734]: pgmap v1233: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:24 vm10.local ceph-mon[51861]: pgmap v1233: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:26 vm08.local ceph-mon[47734]: pgmap v1234: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:26 vm10.local ceph-mon[51861]: pgmap v1234: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:27.091 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:27.092 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:27.119 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:27.119 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:28.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:28 vm08.local ceph-mon[47734]: pgmap v1235: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:28 vm10.local ceph-mon[51861]: pgmap v1235: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:30 vm08.local ceph-mon[47734]: pgmap v1236: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:30 vm10.local ceph-mon[51861]: pgmap v1236: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:32.120 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:32.121 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:32.150 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:32.150 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:32 vm10.local ceph-mon[51861]: pgmap v1237: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:32 vm08.local ceph-mon[47734]: pgmap v1237: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:34 vm08.local ceph-mon[47734]: pgmap v1238: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:34 vm10.local ceph-mon[51861]: pgmap v1238: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:36 vm08.local ceph-mon[47734]: pgmap v1239: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:36 vm10.local ceph-mon[51861]: pgmap v1239: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:37.151 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:37.152 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:37.177 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:37.178 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:38 vm08.local ceph-mon[47734]: pgmap v1240: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:38 vm10.local ceph-mon[51861]: pgmap v1240: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:40.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:40 vm08.local ceph-mon[47734]: pgmap v1241: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:40 vm10.local ceph-mon[51861]: pgmap v1241: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:42.179 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:42.179 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:42.206 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:42.206 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:42.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:42 vm08.local ceph-mon[47734]: pgmap v1242: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:42.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:42 vm10.local ceph-mon[51861]: pgmap v1242: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:34:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:34:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:34:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:34:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:44 vm08.local ceph-mon[47734]: pgmap v1243: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:44 vm10.local ceph-mon[51861]: pgmap v1243: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:47.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:46 vm08.local ceph-mon[47734]: pgmap v1244: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:46 vm10.local ceph-mon[51861]: pgmap v1244: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:47.208 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:47.208 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:47.238 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:47.239 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:48 vm08.local ceph-mon[47734]: pgmap v1245: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:48 vm10.local ceph-mon[51861]: pgmap v1245: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:50.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:49 vm08.local ceph-mon[47734]: pgmap v1246: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:50.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:49 vm10.local ceph-mon[51861]: pgmap v1246: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:52.241 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:52.241 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:52.336 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:52.337 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:52 vm08.local ceph-mon[47734]: pgmap v1247: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:52 vm10.local ceph-mon[51861]: pgmap v1247: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:54 vm08.local ceph-mon[47734]: pgmap v1248: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:54 vm10.local ceph-mon[51861]: pgmap v1248: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:56 vm08.local ceph-mon[47734]: pgmap v1249: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:56 vm10.local ceph-mon[51861]: pgmap v1249: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:34:57.338 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:34:57.339 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:34:57.380 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:34:57.380 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:34:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:34:58 vm08.local ceph-mon[47734]: pgmap v1250: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:34:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:34:58 vm10.local ceph-mon[51861]: pgmap v1250: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:00 vm08.local ceph-mon[47734]: pgmap v1251: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:35:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:35:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:35:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:35:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:00 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:35:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:00 vm10.local ceph-mon[51861]: pgmap v1251: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:35:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:35:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:35:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:35:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:00 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:35:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:35:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:35:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:35:02.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:35:02.382 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:02.382 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:02.439 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:02.440 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:02 vm10.local ceph-mon[51861]: pgmap v1252: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:02 vm08.local ceph-mon[47734]: pgmap v1252: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:04 vm10.local ceph-mon[51861]: pgmap v1253: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:04 vm08.local ceph-mon[47734]: pgmap v1253: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:06 vm08.local ceph-mon[47734]: pgmap v1254: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:06 vm10.local ceph-mon[51861]: pgmap v1254: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:07.441 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:07.442 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:07.477 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:07.477 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:08.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:07 vm08.local ceph-mon[47734]: pgmap v1255: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:07 vm10.local ceph-mon[51861]: pgmap v1255: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:10.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:10 vm08.local ceph-mon[47734]: pgmap v1256: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:10 vm10.local ceph-mon[51861]: pgmap v1256: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:12.479 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:12.479 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:12.505 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:12.505 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:12.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:12 vm08.local ceph-mon[47734]: pgmap v1257: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:12 vm10.local ceph-mon[51861]: pgmap v1257: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:14 vm08.local ceph-mon[47734]: pgmap v1258: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:14 vm10.local ceph-mon[51861]: pgmap v1258: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:16 vm08.local ceph-mon[47734]: pgmap v1259: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:16 vm10.local ceph-mon[51861]: pgmap v1259: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:17.506 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:17.507 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:17.618 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:17.618 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:18 vm08.local ceph-mon[47734]: pgmap v1260: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:18 vm10.local ceph-mon[51861]: pgmap v1260: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:20 vm08.local ceph-mon[47734]: pgmap v1261: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:20 vm10.local ceph-mon[51861]: pgmap v1261: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:22.620 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:22.620 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:22.655 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:22.656 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:22 vm08.local ceph-mon[47734]: pgmap v1262: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:22 vm10.local ceph-mon[51861]: pgmap v1262: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:24 vm08.local ceph-mon[47734]: pgmap v1263: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:24 vm10.local ceph-mon[51861]: pgmap v1263: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:26 vm08.local ceph-mon[47734]: pgmap v1264: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:26 vm10.local ceph-mon[51861]: pgmap v1264: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:27.658 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:27.658 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:27.683 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:27.684 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:28 vm08.local ceph-mon[47734]: pgmap v1265: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:28 vm10.local ceph-mon[51861]: pgmap v1265: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:30 vm08.local ceph-mon[47734]: pgmap v1266: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:30 vm10.local ceph-mon[51861]: pgmap v1266: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:32.686 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:32.686 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:32.713 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:32.713 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:32 vm08.local ceph-mon[47734]: pgmap v1267: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:32 vm10.local ceph-mon[51861]: pgmap v1267: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:34 vm08.local ceph-mon[47734]: pgmap v1268: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:34 vm10.local ceph-mon[51861]: pgmap v1268: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:37.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:36 vm08.local ceph-mon[47734]: pgmap v1269: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:36 vm10.local ceph-mon[51861]: pgmap v1269: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:37.715 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:37.715 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:37.749 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:37.749 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:38 vm08.local ceph-mon[47734]: pgmap v1270: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:38 vm10.local ceph-mon[51861]: pgmap v1270: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:40 vm10.local ceph-mon[51861]: pgmap v1271: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:40 vm08.local ceph-mon[47734]: pgmap v1271: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:42.751 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:42.751 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:42.778 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:42.778 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:42 vm10.local ceph-mon[51861]: pgmap v1272: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:43.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:42 vm08.local ceph-mon[47734]: pgmap v1272: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:35:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:35:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:35:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:35:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:44 vm10.local ceph-mon[51861]: pgmap v1273: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:45.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:44 vm08.local ceph-mon[47734]: pgmap v1273: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:46 vm10.local ceph-mon[51861]: pgmap v1274: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:46 vm08.local ceph-mon[47734]: pgmap v1274: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:47.780 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:47.782 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:47.886 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:47.887 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:48.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:48 vm10.local ceph-mon[51861]: pgmap v1275: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:48 vm08.local ceph-mon[47734]: pgmap v1275: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:50 vm08.local ceph-mon[47734]: pgmap v1276: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:50 vm10.local ceph-mon[51861]: pgmap v1276: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:52.889 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:52.889 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:52.915 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:52.916 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:52 vm08.local ceph-mon[47734]: pgmap v1277: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:52 vm10.local ceph-mon[51861]: pgmap v1277: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:54 vm08.local ceph-mon[47734]: pgmap v1278: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:54 vm10.local ceph-mon[51861]: pgmap v1278: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:56 vm08.local ceph-mon[47734]: pgmap v1279: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:56 vm10.local ceph-mon[51861]: pgmap v1279: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:35:57.917 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:35:57.918 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:35:57.944 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:35:57.945 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:35:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:35:58 vm08.local ceph-mon[47734]: pgmap v1280: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:35:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:35:58 vm10.local ceph-mon[51861]: pgmap v1280: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:00 vm08.local ceph-mon[47734]: pgmap v1281: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:01.124 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:00 vm10.local ceph-mon[51861]: pgmap v1281: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:02.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:36:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:36:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:36:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:36:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:36:02.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:36:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:36:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:36:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:36:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:36:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:36:02.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:36:02.947 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:02.947 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:02.974 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:02.974 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:02 vm08.local ceph-mon[47734]: pgmap v1282: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:02 vm10.local ceph-mon[51861]: pgmap v1282: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:04 vm08.local ceph-mon[47734]: pgmap v1283: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:04 vm10.local ceph-mon[51861]: pgmap v1283: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:07.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:06 vm08.local ceph-mon[47734]: pgmap v1284: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:06 vm10.local ceph-mon[51861]: pgmap v1284: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:07.976 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:07.977 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:08.004 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:08.004 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:09.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:08 vm10.local ceph-mon[51861]: pgmap v1285: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:08 vm08.local ceph-mon[47734]: pgmap v1285: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:10 vm10.local ceph-mon[51861]: pgmap v1286: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:11.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:10 vm08.local ceph-mon[47734]: pgmap v1286: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:13.005 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:13.006 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:13.033 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:13.034 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:12 vm10.local ceph-mon[51861]: pgmap v1287: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:12 vm08.local ceph-mon[47734]: pgmap v1287: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:14 vm10.local ceph-mon[51861]: pgmap v1288: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:14 vm08.local ceph-mon[47734]: pgmap v1288: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:16 vm08.local ceph-mon[47734]: pgmap v1289: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:16 vm10.local ceph-mon[51861]: pgmap v1289: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:18.036 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:18.036 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:18.063 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:18.063 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:18.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:17 vm08.local ceph-mon[47734]: pgmap v1290: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:18.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:17 vm10.local ceph-mon[51861]: pgmap v1290: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:20 vm08.local ceph-mon[47734]: pgmap v1291: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:20 vm10.local ceph-mon[51861]: pgmap v1291: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:22 vm10.local ceph-mon[51861]: pgmap v1292: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:22 vm08.local ceph-mon[47734]: pgmap v1292: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:23.065 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:23.065 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:23.091 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:23.092 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:24 vm10.local ceph-mon[51861]: pgmap v1293: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:24 vm08.local ceph-mon[47734]: pgmap v1293: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:26.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:26 vm08.local ceph-mon[47734]: pgmap v1294: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:26 vm10.local ceph-mon[51861]: pgmap v1294: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:28.093 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:28.094 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:28.120 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:28.121 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:28 vm10.local ceph-mon[51861]: pgmap v1295: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:28 vm08.local ceph-mon[47734]: pgmap v1295: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:30 vm10.local ceph-mon[51861]: pgmap v1296: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:30 vm08.local ceph-mon[47734]: pgmap v1296: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:32 vm10.local ceph-mon[51861]: pgmap v1297: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:32 vm08.local ceph-mon[47734]: pgmap v1297: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:33.122 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:33.122 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:33.150 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:33.150 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:34 vm10.local ceph-mon[51861]: pgmap v1298: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:34 vm08.local ceph-mon[47734]: pgmap v1298: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:37.033 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:36 vm08.local ceph-mon[47734]: pgmap v1299: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:36 vm10.local ceph-mon[51861]: pgmap v1299: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:38.152 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:38.152 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:38.183 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:38.183 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:38 vm08.local ceph-mon[47734]: pgmap v1300: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:38 vm10.local ceph-mon[51861]: pgmap v1300: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:40 vm10.local ceph-mon[51861]: pgmap v1301: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:40 vm08.local ceph-mon[47734]: pgmap v1301: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:42 vm10.local ceph-mon[51861]: pgmap v1302: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:43.185 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:43.185 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:43.212 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:43.213 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:42 vm08.local ceph-mon[47734]: pgmap v1302: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:36:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:36:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:36:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:36:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:44 vm08.local ceph-mon[47734]: pgmap v1303: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:44 vm10.local ceph-mon[51861]: pgmap v1303: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:45 vm08.local ceph-mon[47734]: pgmap v1304: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:46.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:45 vm10.local ceph-mon[51861]: pgmap v1304: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:48.214 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:48.215 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:48.242 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:48.242 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:48 vm08.local ceph-mon[47734]: pgmap v1305: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:48 vm10.local ceph-mon[51861]: pgmap v1305: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:50 vm10.local ceph-mon[51861]: pgmap v1306: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:50 vm08.local ceph-mon[47734]: pgmap v1306: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:52 vm10.local ceph-mon[51861]: pgmap v1307: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:52 vm08.local ceph-mon[47734]: pgmap v1307: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:53.244 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:53.245 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:53.271 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:53.272 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:54 vm10.local ceph-mon[51861]: pgmap v1308: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:54 vm08.local ceph-mon[47734]: pgmap v1308: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:56 vm08.local ceph-mon[47734]: pgmap v1309: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:56 vm10.local ceph-mon[51861]: pgmap v1309: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:36:58.274 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:36:58.274 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:36:58.303 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:36:58.303 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:36:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:36:58 vm10.local ceph-mon[51861]: pgmap v1310: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:36:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:36:58 vm08.local ceph-mon[47734]: pgmap v1310: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:00 vm08.local ceph-mon[47734]: pgmap v1311: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:00 vm10.local ceph-mon[51861]: pgmap v1311: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:01.809 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:37:01.809 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:37:01.809 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:01 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:37:02.019 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:37:02.019 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:37:02.019 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:01 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:37:02.805 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:02 vm08.local ceph-mon[47734]: pgmap v1312: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:02.805 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:02 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:37:02.805 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:02 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:37:02.805 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:02 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:37:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:02 vm10.local ceph-mon[51861]: pgmap v1312: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:02 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:37:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:02 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:37:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:02 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:37:03.305 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:03.306 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:03.366 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:03.367 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:37:04.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:37:04.069 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:37:04.069 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:37:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:04 vm10.local ceph-mon[51861]: pgmap v1313: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:04 vm08.local ceph-mon[47734]: pgmap v1313: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:06.217 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:05 vm08.local ceph-mon[47734]: pgmap v1314: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:06.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:05 vm10.local ceph-mon[51861]: pgmap v1314: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:08.369 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:08.369 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:08.452 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:08.453 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:08 vm10.local ceph-mon[51861]: pgmap v1315: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:08 vm08.local ceph-mon[47734]: pgmap v1315: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:10 vm10.local ceph-mon[51861]: pgmap v1316: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:10 vm08.local ceph-mon[47734]: pgmap v1316: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:12 vm10.local ceph-mon[51861]: pgmap v1317: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:12 vm08.local ceph-mon[47734]: pgmap v1317: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:13.454 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:13.455 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:13.481 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:13.482 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:14.772 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:14 vm10.local ceph-mon[51861]: pgmap v1318: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:14 vm08.local ceph-mon[47734]: pgmap v1318: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:16 vm08.local ceph-mon[47734]: pgmap v1319: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:16 vm10.local ceph-mon[51861]: pgmap v1319: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:18.483 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:18.484 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:18.509 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:18.510 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:18.940 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:18 vm10.local ceph-mon[51861]: pgmap v1320: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:18 vm08.local ceph-mon[47734]: pgmap v1320: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:20 vm08.local ceph-mon[47734]: pgmap v1321: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:20 vm10.local ceph-mon[51861]: pgmap v1321: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:22 vm08.local ceph-mon[47734]: pgmap v1322: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:22 vm10.local ceph-mon[51861]: pgmap v1322: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:23.511 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:23.512 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:23.537 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:23.538 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:24 vm08.local ceph-mon[47734]: pgmap v1323: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:24 vm10.local ceph-mon[51861]: pgmap v1323: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:26.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:26 vm08.local ceph-mon[47734]: pgmap v1324: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:26 vm10.local ceph-mon[51861]: pgmap v1324: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:28.539 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:28.539 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:28.567 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:28.568 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:28 vm08.local ceph-mon[47734]: pgmap v1325: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:28 vm10.local ceph-mon[51861]: pgmap v1325: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:30 vm08.local ceph-mon[47734]: pgmap v1326: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:30 vm10.local ceph-mon[51861]: pgmap v1326: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:32 vm08.local ceph-mon[47734]: pgmap v1327: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:32 vm10.local ceph-mon[51861]: pgmap v1327: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:33.569 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:33.569 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:33.594 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:33.595 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:34.960 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:34 vm10.local ceph-mon[51861]: pgmap v1328: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:34 vm08.local ceph-mon[47734]: pgmap v1328: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:36 vm08.local ceph-mon[47734]: pgmap v1329: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:36 vm10.local ceph-mon[51861]: pgmap v1329: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:38.596 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:38.596 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:38.622 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:38.622 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:38 vm08.local ceph-mon[47734]: pgmap v1330: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:38 vm10.local ceph-mon[51861]: pgmap v1330: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:40 vm08.local ceph-mon[47734]: pgmap v1331: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:40 vm10.local ceph-mon[51861]: pgmap v1331: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:42 vm08.local ceph-mon[47734]: pgmap v1332: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:42 vm10.local ceph-mon[51861]: pgmap v1332: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:43.623 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:43.624 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:43.651 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:43.651 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:37:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:37:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:37:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:37:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:44 vm08.local ceph-mon[47734]: pgmap v1333: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:44 vm10.local ceph-mon[51861]: pgmap v1333: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:46 vm10.local ceph-mon[51861]: pgmap v1334: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:46 vm08.local ceph-mon[47734]: pgmap v1334: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:48.653 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:48.653 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:48.679 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:48.679 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:48 vm10.local ceph-mon[51861]: pgmap v1335: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:48 vm08.local ceph-mon[47734]: pgmap v1335: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:50 vm10.local ceph-mon[51861]: pgmap v1336: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:50 vm08.local ceph-mon[47734]: pgmap v1336: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:52 vm10.local ceph-mon[51861]: pgmap v1337: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:52 vm08.local ceph-mon[47734]: pgmap v1337: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:53.681 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:53.681 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:53.707 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:53.708 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:54 vm10.local ceph-mon[51861]: pgmap v1338: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:54 vm08.local ceph-mon[47734]: pgmap v1338: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:56 vm10.local ceph-mon[51861]: pgmap v1339: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:56 vm08.local ceph-mon[47734]: pgmap v1339: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:37:58.709 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:37:58.709 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:37:58.735 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:37:58.736 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:37:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:37:58 vm10.local ceph-mon[51861]: pgmap v1340: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:37:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:37:58 vm08.local ceph-mon[47734]: pgmap v1340: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:00 vm10.local ceph-mon[51861]: pgmap v1341: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:00 vm08.local ceph-mon[47734]: pgmap v1341: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:02 vm10.local ceph-mon[51861]: pgmap v1342: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:02 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:38:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:02 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:38:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:02 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:38:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:02 vm08.local ceph-mon[47734]: pgmap v1342: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:02 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:38:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:02 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:38:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:02 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:38:03.737 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:03.738 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:03.763 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:03.764 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:38:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:38:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:04 vm10.local ceph-mon[51861]: pgmap v1343: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:38:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:38:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:04 vm08.local ceph-mon[47734]: pgmap v1343: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:06.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:06 vm08.local ceph-mon[47734]: pgmap v1344: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:06 vm10.local ceph-mon[51861]: pgmap v1344: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:08.765 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:08.766 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:08.795 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:08.796 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:08.850 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:08 vm08.local ceph-mon[47734]: pgmap v1345: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:08 vm10.local ceph-mon[51861]: pgmap v1345: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:10 vm10.local ceph-mon[51861]: pgmap v1346: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:10 vm08.local ceph-mon[47734]: pgmap v1346: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:12 vm10.local ceph-mon[51861]: pgmap v1347: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:38:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:12 vm08.local ceph-mon[47734]: pgmap v1347: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:38:13.797 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:13.797 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:13.825 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:13.826 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:14 vm10.local ceph-mon[51861]: pgmap v1348: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:38:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:14 vm08.local ceph-mon[47734]: pgmap v1348: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:38:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:16 vm08.local ceph-mon[47734]: pgmap v1349: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:16 vm10.local ceph-mon[51861]: pgmap v1349: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:18.827 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:18.828 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:18.854 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:18.855 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:18 vm10.local ceph-mon[51861]: pgmap v1350: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:38:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:18 vm08.local ceph-mon[47734]: pgmap v1350: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:38:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:20 vm08.local ceph-mon[47734]: pgmap v1351: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:20 vm10.local ceph-mon[51861]: pgmap v1351: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:22 vm08.local ceph-mon[47734]: pgmap v1352: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:22 vm10.local ceph-mon[51861]: pgmap v1352: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:23.856 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:23.857 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:23.883 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:23.884 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:24 vm08.local ceph-mon[47734]: pgmap v1353: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:24 vm10.local ceph-mon[51861]: pgmap v1353: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:26 vm08.local ceph-mon[47734]: pgmap v1354: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:26 vm10.local ceph-mon[51861]: pgmap v1354: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:28.885 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:28.886 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:28.913 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:28.914 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:28 vm08.local ceph-mon[47734]: pgmap v1355: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:29.162 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:28 vm10.local ceph-mon[51861]: pgmap v1355: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:30 vm10.local ceph-mon[51861]: pgmap v1356: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:30 vm08.local ceph-mon[47734]: pgmap v1356: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:32 vm10.local ceph-mon[51861]: pgmap v1357: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:32 vm08.local ceph-mon[47734]: pgmap v1357: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:33.915 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:33.916 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:33.943 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:33.943 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:34 vm10.local ceph-mon[51861]: pgmap v1358: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:34 vm08.local ceph-mon[47734]: pgmap v1358: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:36 vm10.local ceph-mon[51861]: pgmap v1359: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:36 vm08.local ceph-mon[47734]: pgmap v1359: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:38.945 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:38.945 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:38.971 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:38.972 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:38 vm10.local ceph-mon[51861]: pgmap v1360: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:38 vm08.local ceph-mon[47734]: pgmap v1360: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:40 vm10.local ceph-mon[51861]: pgmap v1361: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:40 vm08.local ceph-mon[47734]: pgmap v1361: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:42 vm10.local ceph-mon[51861]: pgmap v1362: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:42 vm08.local ceph-mon[47734]: pgmap v1362: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:43.973 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:43.974 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:44.001 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:44.002 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:38:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:38:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:38:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:38:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:44 vm10.local ceph-mon[51861]: pgmap v1363: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:44 vm08.local ceph-mon[47734]: pgmap v1363: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:46 vm10.local ceph-mon[51861]: pgmap v1364: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:46 vm08.local ceph-mon[47734]: pgmap v1364: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:49.003 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:49.004 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:49.028 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:49.029 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:48 vm10.local ceph-mon[51861]: pgmap v1365: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:48 vm08.local ceph-mon[47734]: pgmap v1365: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:50 vm08.local ceph-mon[47734]: pgmap v1366: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:50 vm10.local ceph-mon[51861]: pgmap v1366: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:52 vm08.local ceph-mon[47734]: pgmap v1367: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:52 vm10.local ceph-mon[51861]: pgmap v1367: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:54.030 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:54.031 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:54.056 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:54.056 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:38:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:54 vm08.local ceph-mon[47734]: pgmap v1368: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:54 vm10.local ceph-mon[51861]: pgmap v1368: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:56.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:55 vm08.local ceph-mon[47734]: pgmap v1369: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:56.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:55 vm10.local ceph-mon[51861]: pgmap v1369: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:38:58.855 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:38:58 vm08.local ceph-mon[47734]: pgmap v1370: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:38:58 vm10.local ceph-mon[51861]: pgmap v1370: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:38:59.058 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:38:59.059 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:38:59.126 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:38:59.127 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:00 vm10.local ceph-mon[51861]: pgmap v1371: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:00 vm08.local ceph-mon[47734]: pgmap v1371: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:02 vm10.local ceph-mon[51861]: pgmap v1372: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:02 vm08.local ceph-mon[47734]: pgmap v1372: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:39:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:39:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:39:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:39:03.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:39:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:39:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:39:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:39:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:39:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:39:04.128 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:04.129 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:04.154 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:04.155 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:04 vm10.local ceph-mon[51861]: pgmap v1373: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:04 vm08.local ceph-mon[47734]: pgmap v1373: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:06 vm08.local ceph-mon[47734]: pgmap v1374: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:06 vm10.local ceph-mon[51861]: pgmap v1374: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:08 vm10.local ceph-mon[51861]: pgmap v1375: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:08 vm08.local ceph-mon[47734]: pgmap v1375: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:09.156 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:09.157 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:09.183 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:09.184 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:10 vm10.local ceph-mon[51861]: pgmap v1376: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:10 vm08.local ceph-mon[47734]: pgmap v1376: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:12 vm10.local ceph-mon[51861]: pgmap v1377: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:12 vm08.local ceph-mon[47734]: pgmap v1377: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:14.185 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:14.186 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:14.211 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:14.212 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:14 vm10.local ceph-mon[51861]: pgmap v1378: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:14 vm08.local ceph-mon[47734]: pgmap v1378: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:16 vm08.local ceph-mon[47734]: pgmap v1379: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:16 vm10.local ceph-mon[51861]: pgmap v1379: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:18 vm10.local ceph-mon[51861]: pgmap v1380: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:19.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:18 vm08.local ceph-mon[47734]: pgmap v1380: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:19.213 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:19.214 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:19.241 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:19.242 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:20 vm10.local ceph-mon[51861]: pgmap v1381: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:20 vm08.local ceph-mon[47734]: pgmap v1381: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:22 vm08.local ceph-mon[47734]: pgmap v1382: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:22 vm10.local ceph-mon[51861]: pgmap v1382: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:24.243 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:24.243 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:24.268 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:24.269 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:24 vm08.local ceph-mon[47734]: pgmap v1383: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:24 vm10.local ceph-mon[51861]: pgmap v1383: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:26 vm08.local ceph-mon[47734]: pgmap v1384: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:26 vm10.local ceph-mon[51861]: pgmap v1384: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:28 vm08.local ceph-mon[47734]: pgmap v1385: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:28 vm10.local ceph-mon[51861]: pgmap v1385: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:29.270 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:29.271 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:29.298 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:29.299 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:30 vm08.local ceph-mon[47734]: pgmap v1386: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:30 vm10.local ceph-mon[51861]: pgmap v1386: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:32 vm08.local ceph-mon[47734]: pgmap v1387: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:32 vm10.local ceph-mon[51861]: pgmap v1387: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:34.300 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:34.301 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:34.326 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:34.327 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:34 vm08.local ceph-mon[47734]: pgmap v1388: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:34 vm10.local ceph-mon[51861]: pgmap v1388: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:36 vm08.local ceph-mon[47734]: pgmap v1389: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:36 vm10.local ceph-mon[51861]: pgmap v1389: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:38 vm08.local ceph-mon[47734]: pgmap v1390: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:38 vm10.local ceph-mon[51861]: pgmap v1390: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:39.328 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:39.329 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:39.354 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:39.355 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:40 vm08.local ceph-mon[47734]: pgmap v1391: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:40 vm10.local ceph-mon[51861]: pgmap v1391: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:42 vm08.local ceph-mon[47734]: pgmap v1392: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:42 vm10.local ceph-mon[51861]: pgmap v1392: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:39:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:39:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:39:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:39:44.356 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:44.357 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:44.382 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:44.383 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:44 vm08.local ceph-mon[47734]: pgmap v1393: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:44 vm10.local ceph-mon[51861]: pgmap v1393: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:47.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:46 vm08.local ceph-mon[47734]: pgmap v1394: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:46 vm10.local ceph-mon[51861]: pgmap v1394: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:48 vm08.local ceph-mon[47734]: pgmap v1395: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:48 vm10.local ceph-mon[51861]: pgmap v1395: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:49.384 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:49.384 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:49.410 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:49.410 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:50 vm08.local ceph-mon[47734]: pgmap v1396: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:50 vm10.local ceph-mon[51861]: pgmap v1396: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:52 vm08.local ceph-mon[47734]: pgmap v1397: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:52 vm10.local ceph-mon[51861]: pgmap v1397: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:54.412 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:54.412 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:54.439 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:54.439 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:39:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:54 vm08.local ceph-mon[47734]: pgmap v1398: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:54 vm10.local ceph-mon[51861]: pgmap v1398: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:57.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:56 vm08.local ceph-mon[47734]: pgmap v1399: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:56 vm10.local ceph-mon[51861]: pgmap v1399: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:39:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:39:58 vm08.local ceph-mon[47734]: pgmap v1400: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:39:58 vm10.local ceph-mon[51861]: pgmap v1400: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:39:59.441 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:39:59.441 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:39:59.467 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:39:59.468 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:00 vm08.local ceph-mon[47734]: pgmap v1401: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T14:40:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:00 vm10.local ceph-mon[51861]: pgmap v1401: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:01.176 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T14:40:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:02 vm08.local ceph-mon[47734]: pgmap v1402: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:02 vm10.local ceph-mon[51861]: pgmap v1402: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:40:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:40:04.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:40:04.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:40:04.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:40:04.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:40:04.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:40:04.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:40:04.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:40:04.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:40:04.469 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:04.470 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:04.496 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:04.496 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:04 vm10.local ceph-mon[51861]: pgmap v1403: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:04 vm08.local ceph-mon[47734]: pgmap v1403: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:07.066 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:06 vm08.local ceph-mon[47734]: pgmap v1404: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:06 vm10.local ceph-mon[51861]: pgmap v1404: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:09.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:08 vm10.local ceph-mon[51861]: pgmap v1405: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:08 vm08.local ceph-mon[47734]: pgmap v1405: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:09.498 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:09.499 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:09.525 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:09.526 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:10 vm10.local ceph-mon[51861]: pgmap v1406: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:10 vm08.local ceph-mon[47734]: pgmap v1406: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:12 vm10.local ceph-mon[51861]: pgmap v1407: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:12 vm08.local ceph-mon[47734]: pgmap v1407: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:14.527 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:14.528 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:14.555 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:14.555 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:14 vm10.local ceph-mon[51861]: pgmap v1408: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:14 vm08.local ceph-mon[47734]: pgmap v1408: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:16 vm10.local ceph-mon[51861]: pgmap v1409: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:17.220 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:16 vm08.local ceph-mon[47734]: pgmap v1409: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:18 vm10.local ceph-mon[51861]: pgmap v1410: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:18 vm08.local ceph-mon[47734]: pgmap v1410: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:19.557 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:19.557 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:19.585 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:19.586 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:20 vm10.local ceph-mon[51861]: pgmap v1411: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:20 vm08.local ceph-mon[47734]: pgmap v1411: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:22 vm10.local ceph-mon[51861]: pgmap v1412: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:22 vm08.local ceph-mon[47734]: pgmap v1412: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:24.587 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:24.587 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:24.612 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:24.613 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:24 vm10.local ceph-mon[51861]: pgmap v1413: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:24 vm08.local ceph-mon[47734]: pgmap v1413: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:26 vm10.local ceph-mon[51861]: pgmap v1414: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:26 vm08.local ceph-mon[47734]: pgmap v1414: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:28 vm10.local ceph-mon[51861]: pgmap v1415: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:28 vm08.local ceph-mon[47734]: pgmap v1415: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:29.614 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:29.615 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:29.640 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:29.640 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:30 vm08.local ceph-mon[47734]: pgmap v1416: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:30 vm10.local ceph-mon[51861]: pgmap v1416: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:32 vm08.local ceph-mon[47734]: pgmap v1417: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:32 vm10.local ceph-mon[51861]: pgmap v1417: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:34.641 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:34.642 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:34.700 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:34.701 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:34 vm08.local ceph-mon[47734]: pgmap v1418: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:34 vm10.local ceph-mon[51861]: pgmap v1418: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:37.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:36 vm08.local ceph-mon[47734]: pgmap v1419: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:36 vm10.local ceph-mon[51861]: pgmap v1419: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:39.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:38 vm08.local ceph-mon[47734]: pgmap v1420: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:38 vm10.local ceph-mon[51861]: pgmap v1420: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:39.702 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:39.702 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:39.730 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:39.731 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:40 vm10.local ceph-mon[51861]: pgmap v1421: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:40 vm08.local ceph-mon[47734]: pgmap v1421: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:42 vm10.local ceph-mon[51861]: pgmap v1422: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:42 vm08.local ceph-mon[47734]: pgmap v1422: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:40:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:40:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:40:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:40:44.732 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:44.732 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:44.761 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:44.761 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:44 vm08.local ceph-mon[47734]: pgmap v1423: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:44 vm10.local ceph-mon[51861]: pgmap v1423: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:46 vm08.local ceph-mon[47734]: pgmap v1424: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:46 vm10.local ceph-mon[51861]: pgmap v1424: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:48 vm08.local ceph-mon[47734]: pgmap v1425: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:48 vm10.local ceph-mon[51861]: pgmap v1425: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:49.762 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:49.763 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:49.789 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:49.790 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:50 vm08.local ceph-mon[47734]: pgmap v1426: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:50 vm10.local ceph-mon[51861]: pgmap v1426: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:52 vm08.local ceph-mon[47734]: pgmap v1427: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:52 vm10.local ceph-mon[51861]: pgmap v1427: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:54.792 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:54.792 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:54.818 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:54.819 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:40:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:54 vm08.local ceph-mon[47734]: pgmap v1428: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:54 vm10.local ceph-mon[51861]: pgmap v1428: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:56.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:55 vm08.local ceph-mon[47734]: pgmap v1429: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:56.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:55 vm10.local ceph-mon[51861]: pgmap v1429: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:40:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:40:58 vm10.local ceph-mon[51861]: pgmap v1430: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:40:58 vm08.local ceph-mon[47734]: pgmap v1430: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:40:59.821 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:40:59.821 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:40:59.847 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:40:59.848 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:00 vm10.local ceph-mon[51861]: pgmap v1431: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:00 vm08.local ceph-mon[47734]: pgmap v1431: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:02 vm08.local ceph-mon[47734]: pgmap v1432: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:02 vm10.local ceph-mon[51861]: pgmap v1432: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:04.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:41:04.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:41:04.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:41:04.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:41:04.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:41:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:41:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:41:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:41:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:41:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:41:04.849 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:04.850 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:04.875 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:04.876 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:04 vm08.local ceph-mon[47734]: pgmap v1433: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:04 vm10.local ceph-mon[51861]: pgmap v1433: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:06.208 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:05 vm08.local ceph-mon[47734]: pgmap v1434: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:06.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:05 vm10.local ceph-mon[51861]: pgmap v1434: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:09.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:08 vm08.local ceph-mon[47734]: pgmap v1435: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:09.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:08 vm10.local ceph-mon[51861]: pgmap v1435: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:09.877 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:09.877 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:09.903 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:09.904 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:10 vm10.local ceph-mon[51861]: pgmap v1436: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:10 vm08.local ceph-mon[47734]: pgmap v1436: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:12 vm10.local ceph-mon[51861]: pgmap v1437: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:12 vm08.local ceph-mon[47734]: pgmap v1437: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:14.905 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:14.906 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:14.931 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:14.932 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:14 vm10.local ceph-mon[51861]: pgmap v1438: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:14 vm08.local ceph-mon[47734]: pgmap v1438: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:16 vm10.local ceph-mon[51861]: pgmap v1439: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:16 vm08.local ceph-mon[47734]: pgmap v1439: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:18 vm10.local ceph-mon[51861]: pgmap v1440: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:18 vm08.local ceph-mon[47734]: pgmap v1440: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:19.933 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:19.933 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:19.959 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:19.959 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:20 vm10.local ceph-mon[51861]: pgmap v1441: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:20 vm08.local ceph-mon[47734]: pgmap v1441: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:22 vm10.local ceph-mon[51861]: pgmap v1442: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:22 vm08.local ceph-mon[47734]: pgmap v1442: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:24.961 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:24.961 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:24.987 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:24.988 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:24 vm10.local ceph-mon[51861]: pgmap v1443: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:24 vm08.local ceph-mon[47734]: pgmap v1443: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:26 vm10.local ceph-mon[51861]: pgmap v1444: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:26 vm08.local ceph-mon[47734]: pgmap v1444: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:28 vm10.local ceph-mon[51861]: pgmap v1445: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:28 vm08.local ceph-mon[47734]: pgmap v1445: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:29.989 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:29.989 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:30.015 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:30.016 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:30 vm10.local ceph-mon[51861]: pgmap v1446: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:30 vm08.local ceph-mon[47734]: pgmap v1446: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:32 vm10.local ceph-mon[51861]: pgmap v1447: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:32 vm08.local ceph-mon[47734]: pgmap v1447: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:35.017 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:35.018 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:35.043 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:35.044 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:34 vm10.local ceph-mon[51861]: pgmap v1448: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:34 vm08.local ceph-mon[47734]: pgmap v1448: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:36 vm10.local ceph-mon[51861]: pgmap v1449: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:36 vm08.local ceph-mon[47734]: pgmap v1449: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:38 vm10.local ceph-mon[51861]: pgmap v1450: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:38 vm08.local ceph-mon[47734]: pgmap v1450: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:40.045 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:40.046 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:40.071 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:40.071 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:40 vm10.local ceph-mon[51861]: pgmap v1451: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:40 vm08.local ceph-mon[47734]: pgmap v1451: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:42 vm10.local ceph-mon[51861]: pgmap v1452: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:42 vm08.local ceph-mon[47734]: pgmap v1452: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:41:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:41:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:41:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:41:45.072 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:45.073 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:45.098 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:45.099 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:44 vm10.local ceph-mon[51861]: pgmap v1453: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:44 vm08.local ceph-mon[47734]: pgmap v1453: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:46 vm08.local ceph-mon[47734]: pgmap v1454: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:46 vm10.local ceph-mon[51861]: pgmap v1454: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:48 vm08.local ceph-mon[47734]: pgmap v1455: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:48 vm10.local ceph-mon[51861]: pgmap v1455: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:50.100 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:50.101 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:50.125 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:50.126 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:50 vm08.local ceph-mon[47734]: pgmap v1456: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:50 vm10.local ceph-mon[51861]: pgmap v1456: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:52 vm08.local ceph-mon[47734]: pgmap v1457: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:52 vm10.local ceph-mon[51861]: pgmap v1457: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:55.127 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:41:55.128 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:41:55.153 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:41:55.154 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:41:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:54 vm08.local ceph-mon[47734]: pgmap v1458: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:54 vm10.local ceph-mon[51861]: pgmap v1458: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:56 vm08.local ceph-mon[47734]: pgmap v1459: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:56 vm10.local ceph-mon[51861]: pgmap v1459: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:41:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:41:58 vm08.local ceph-mon[47734]: pgmap v1460: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:41:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:41:58 vm10.local ceph-mon[51861]: pgmap v1460: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:00.155 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:00.156 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:00.181 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:00.182 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:00 vm08.local ceph-mon[47734]: pgmap v1461: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:00 vm10.local ceph-mon[51861]: pgmap v1461: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:02 vm08.local ceph-mon[47734]: pgmap v1462: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:02 vm10.local ceph-mon[51861]: pgmap v1462: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:04.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:03 vm08.local ceph-mon[47734]: pgmap v1463: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:04.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:42:04.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:42:04.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:03 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:42:04.309 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:03 vm10.local ceph-mon[51861]: pgmap v1463: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:04.309 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:42:04.309 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:42:04.309 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:03 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:42:05.183 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:05.183 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:05.210 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:05.210 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:05 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:42:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:05 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:42:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:05 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:42:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:05 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:42:06.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:06 vm10.local ceph-mon[51861]: pgmap v1464: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:06.458 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:06 vm08.local ceph-mon[47734]: pgmap v1464: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:09.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:08 vm10.local ceph-mon[51861]: pgmap v1465: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:08 vm08.local ceph-mon[47734]: pgmap v1465: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:10.211 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:10.212 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:10.237 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:10.237 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:10 vm10.local ceph-mon[51861]: pgmap v1466: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:10 vm08.local ceph-mon[47734]: pgmap v1466: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:12 vm10.local ceph-mon[51861]: pgmap v1467: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:12 vm08.local ceph-mon[47734]: pgmap v1467: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:14 vm10.local ceph-mon[51861]: pgmap v1468: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:15.238 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:15.239 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:15.266 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:15.266 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:14 vm08.local ceph-mon[47734]: pgmap v1468: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:16 vm10.local ceph-mon[51861]: pgmap v1469: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:16 vm08.local ceph-mon[47734]: pgmap v1469: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:18 vm10.local ceph-mon[51861]: pgmap v1470: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:18 vm08.local ceph-mon[47734]: pgmap v1470: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:20.267 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:20.268 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:20.293 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:20.294 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:20 vm10.local ceph-mon[51861]: pgmap v1471: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:20 vm08.local ceph-mon[47734]: pgmap v1471: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:22 vm10.local ceph-mon[51861]: pgmap v1472: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:22 vm08.local ceph-mon[47734]: pgmap v1472: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:24 vm10.local ceph-mon[51861]: pgmap v1473: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:25.295 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:25.295 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:24 vm08.local ceph-mon[47734]: pgmap v1473: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:25.320 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:25.320 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:26 vm08.local ceph-mon[47734]: pgmap v1474: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:27 vm10.local ceph-mon[51861]: pgmap v1474: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:28.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:28 vm08.local ceph-mon[47734]: pgmap v1475: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:28.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:28 vm10.local ceph-mon[51861]: pgmap v1475: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:30.322 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:30.322 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:30.394 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:30.394 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:30 vm08.local ceph-mon[47734]: pgmap v1476: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:30 vm10.local ceph-mon[51861]: pgmap v1476: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:32 vm10.local ceph-mon[51861]: pgmap v1477: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:32 vm08.local ceph-mon[47734]: pgmap v1477: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:34 vm10.local ceph-mon[51861]: pgmap v1478: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:34 vm08.local ceph-mon[47734]: pgmap v1478: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:35.395 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:35.396 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:35.421 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:35.422 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:36 vm10.local ceph-mon[51861]: pgmap v1479: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:36 vm08.local ceph-mon[47734]: pgmap v1479: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:38 vm10.local ceph-mon[51861]: pgmap v1480: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:38 vm08.local ceph-mon[47734]: pgmap v1480: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:40.423 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:40.424 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:40.450 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:40.450 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:40 vm10.local ceph-mon[51861]: pgmap v1481: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:40 vm08.local ceph-mon[47734]: pgmap v1481: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:42 vm10.local ceph-mon[51861]: pgmap v1482: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:42 vm08.local ceph-mon[47734]: pgmap v1482: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:42:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:42:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:42:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:42:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:44 vm10.local ceph-mon[51861]: pgmap v1483: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:44 vm08.local ceph-mon[47734]: pgmap v1483: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:45.452 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:45.452 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:45.479 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:45.479 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:46 vm10.local ceph-mon[51861]: pgmap v1484: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:46 vm08.local ceph-mon[47734]: pgmap v1484: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:48 vm10.local ceph-mon[51861]: pgmap v1485: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:48 vm08.local ceph-mon[47734]: pgmap v1485: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:50.481 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:50.481 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:50.507 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:50.508 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:50 vm10.local ceph-mon[51861]: pgmap v1486: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:50 vm08.local ceph-mon[47734]: pgmap v1486: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:52 vm10.local ceph-mon[51861]: pgmap v1487: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:52 vm08.local ceph-mon[47734]: pgmap v1487: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:54 vm10.local ceph-mon[51861]: pgmap v1488: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:54 vm08.local ceph-mon[47734]: pgmap v1488: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:55.509 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:42:55.509 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:42:55.538 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:42:55.538 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:42:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:56 vm10.local ceph-mon[51861]: pgmap v1489: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:56 vm08.local ceph-mon[47734]: pgmap v1489: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:42:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:42:58 vm10.local ceph-mon[51861]: pgmap v1490: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:42:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:42:58 vm08.local ceph-mon[47734]: pgmap v1490: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:00.539 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:00.540 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:00.567 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:00.567 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:00 vm10.local ceph-mon[51861]: pgmap v1491: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:00 vm08.local ceph-mon[47734]: pgmap v1491: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:02 vm10.local ceph-mon[51861]: pgmap v1492: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:02 vm08.local ceph-mon[47734]: pgmap v1492: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:04 vm08.local ceph-mon[47734]: pgmap v1493: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:43:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:43:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:43:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:43:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:43:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:04 vm10.local ceph-mon[51861]: pgmap v1493: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:43:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:43:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:43:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:43:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:43:05.568 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:05.569 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:05.599 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:05.599 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:06 vm08.local ceph-mon[47734]: pgmap v1494: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:06 vm10.local ceph-mon[51861]: pgmap v1494: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:08 vm08.local ceph-mon[47734]: pgmap v1495: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:08 vm10.local ceph-mon[51861]: pgmap v1495: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:10.601 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:10.601 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:10.627 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:10.627 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:10 vm08.local ceph-mon[47734]: pgmap v1496: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:10 vm10.local ceph-mon[51861]: pgmap v1496: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:12 vm08.local ceph-mon[47734]: pgmap v1497: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:12 vm10.local ceph-mon[51861]: pgmap v1497: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:14 vm08.local ceph-mon[47734]: pgmap v1498: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:14 vm10.local ceph-mon[51861]: pgmap v1498: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:15.629 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:15.629 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:15.655 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:15.656 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:16 vm08.local ceph-mon[47734]: pgmap v1499: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:16 vm10.local ceph-mon[51861]: pgmap v1499: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:18 vm08.local ceph-mon[47734]: pgmap v1500: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:18 vm10.local ceph-mon[51861]: pgmap v1500: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:20.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:20 vm08.local ceph-mon[47734]: pgmap v1501: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:20.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:20 vm10.local ceph-mon[51861]: pgmap v1501: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:20.657 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:20.658 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:20.683 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:20.684 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:22 vm08.local ceph-mon[47734]: pgmap v1502: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:22 vm10.local ceph-mon[51861]: pgmap v1502: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:24 vm08.local ceph-mon[47734]: pgmap v1503: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:24 vm10.local ceph-mon[51861]: pgmap v1503: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:25.685 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:25.685 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:25.710 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:25.710 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:26.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:26 vm08.local ceph-mon[47734]: pgmap v1504: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:26 vm10.local ceph-mon[51861]: pgmap v1504: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:28 vm08.local ceph-mon[47734]: pgmap v1505: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:28 vm10.local ceph-mon[51861]: pgmap v1505: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:30 vm08.local ceph-mon[47734]: pgmap v1506: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:30 vm10.local ceph-mon[51861]: pgmap v1506: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:30.711 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:30.712 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:30.737 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:30.738 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:32 vm08.local ceph-mon[47734]: pgmap v1507: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:32 vm10.local ceph-mon[51861]: pgmap v1507: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:35.210 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:34 vm10.local ceph-mon[51861]: pgmap v1508: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:34 vm08.local ceph-mon[47734]: pgmap v1508: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:35.739 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:35.740 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:35.766 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:35.766 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:36.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:36 vm08.local ceph-mon[47734]: pgmap v1509: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:36 vm10.local ceph-mon[51861]: pgmap v1509: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:38 vm08.local ceph-mon[47734]: pgmap v1510: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:38 vm10.local ceph-mon[51861]: pgmap v1510: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:40.768 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:40.768 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:40.793 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:40.794 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:40 vm08.local ceph-mon[47734]: pgmap v1511: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:40 vm10.local ceph-mon[51861]: pgmap v1511: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:42 vm08.local ceph-mon[47734]: pgmap v1512: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:42 vm10.local ceph-mon[51861]: pgmap v1512: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:43:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:43:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:43:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:43:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:44 vm08.local ceph-mon[47734]: pgmap v1513: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:44 vm10.local ceph-mon[51861]: pgmap v1513: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:45.795 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:45.795 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:45.822 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:45.823 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:46 vm08.local ceph-mon[47734]: pgmap v1514: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:46 vm10.local ceph-mon[51861]: pgmap v1514: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:48.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:48 vm10.local ceph-mon[51861]: pgmap v1515: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:48 vm08.local ceph-mon[47734]: pgmap v1515: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:50 vm08.local ceph-mon[47734]: pgmap v1516: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:50.824 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:50.824 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:50.850 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:50.851 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:50 vm10.local ceph-mon[51861]: pgmap v1516: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:52 vm08.local ceph-mon[47734]: pgmap v1517: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:52 vm10.local ceph-mon[51861]: pgmap v1517: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:54 vm08.local ceph-mon[47734]: pgmap v1518: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:54 vm10.local ceph-mon[51861]: pgmap v1518: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:55.852 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:43:55.853 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:43:55.880 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:43:55.880 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:43:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:56 vm08.local ceph-mon[47734]: pgmap v1519: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:56 vm10.local ceph-mon[51861]: pgmap v1519: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:43:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:43:58 vm08.local ceph-mon[47734]: pgmap v1520: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:43:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:43:58 vm10.local ceph-mon[51861]: pgmap v1520: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:00.882 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:00.882 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:00.909 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:00.909 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:00 vm08.local ceph-mon[47734]: pgmap v1521: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:00 vm10.local ceph-mon[51861]: pgmap v1521: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:02 vm08.local ceph-mon[47734]: pgmap v1522: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:02 vm10.local ceph-mon[51861]: pgmap v1522: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:04 vm08.local ceph-mon[47734]: pgmap v1523: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:05.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:44:05.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:44:05.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:44:05.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:44:05.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:44:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:04 vm10.local ceph-mon[51861]: pgmap v1523: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:44:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:44:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:44:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:44:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:44:05.910 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:05.911 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:05.936 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:05.937 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:06 vm08.local ceph-mon[47734]: pgmap v1524: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:06 vm10.local ceph-mon[51861]: pgmap v1524: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:08 vm08.local ceph-mon[47734]: pgmap v1525: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:08 vm10.local ceph-mon[51861]: pgmap v1525: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:10.938 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:10.938 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:10.964 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:10.964 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:10 vm08.local ceph-mon[47734]: pgmap v1526: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:10 vm10.local ceph-mon[51861]: pgmap v1526: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:12 vm08.local ceph-mon[47734]: pgmap v1527: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:12 vm10.local ceph-mon[51861]: pgmap v1527: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:14 vm08.local ceph-mon[47734]: pgmap v1528: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:14 vm10.local ceph-mon[51861]: pgmap v1528: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:15.965 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:15.966 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:15.991 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:15.991 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:16 vm08.local ceph-mon[47734]: pgmap v1529: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:16 vm10.local ceph-mon[51861]: pgmap v1529: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:18 vm08.local ceph-mon[47734]: pgmap v1530: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:18 vm10.local ceph-mon[51861]: pgmap v1530: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:20.992 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:20.993 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:21.020 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:21.021 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:20 vm08.local ceph-mon[47734]: pgmap v1531: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:20 vm10.local ceph-mon[51861]: pgmap v1531: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:22 vm08.local ceph-mon[47734]: pgmap v1532: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:22 vm10.local ceph-mon[51861]: pgmap v1532: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:24 vm08.local ceph-mon[47734]: pgmap v1533: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:24 vm10.local ceph-mon[51861]: pgmap v1533: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:26.022 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:26.022 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:26.048 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:26.049 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:26 vm08.local ceph-mon[47734]: pgmap v1534: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:26 vm10.local ceph-mon[51861]: pgmap v1534: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:28 vm08.local ceph-mon[47734]: pgmap v1535: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:28 vm10.local ceph-mon[51861]: pgmap v1535: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:30 vm08.local ceph-mon[47734]: pgmap v1536: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:30 vm10.local ceph-mon[51861]: pgmap v1536: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:31.050 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:31.050 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:31.078 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:31.078 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:32 vm08.local ceph-mon[47734]: pgmap v1537: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:32 vm10.local ceph-mon[51861]: pgmap v1537: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:35.210 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:34 vm10.local ceph-mon[51861]: pgmap v1538: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:34 vm08.local ceph-mon[47734]: pgmap v1538: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:36.079 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:36.080 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:36.106 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:36.107 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:36 vm08.local ceph-mon[47734]: pgmap v1539: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:36 vm10.local ceph-mon[51861]: pgmap v1539: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:38 vm08.local ceph-mon[47734]: pgmap v1540: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:38 vm10.local ceph-mon[51861]: pgmap v1540: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:41.108 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:41.108 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:41.134 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:41.135 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:40 vm08.local ceph-mon[47734]: pgmap v1541: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:40 vm10.local ceph-mon[51861]: pgmap v1541: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:42 vm08.local ceph-mon[47734]: pgmap v1542: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:42 vm10.local ceph-mon[51861]: pgmap v1542: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:44:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:44:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:44:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:44:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:44 vm08.local ceph-mon[47734]: pgmap v1543: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:44 vm10.local ceph-mon[51861]: pgmap v1543: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:46.136 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:46.137 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:46.162 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:46.162 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:46 vm08.local ceph-mon[47734]: pgmap v1544: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:46 vm10.local ceph-mon[51861]: pgmap v1544: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:48.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:48 vm08.local ceph-mon[47734]: pgmap v1545: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:48.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:48 vm10.local ceph-mon[51861]: pgmap v1545: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:50 vm10.local ceph-mon[51861]: pgmap v1546: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:50 vm08.local ceph-mon[47734]: pgmap v1546: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:51.164 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:51.164 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:51.189 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:51.189 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:52 vm08.local ceph-mon[47734]: pgmap v1547: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:52 vm10.local ceph-mon[51861]: pgmap v1547: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:54 vm08.local ceph-mon[47734]: pgmap v1548: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:54 vm10.local ceph-mon[51861]: pgmap v1548: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:56.190 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:44:56.191 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:44:56.216 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:44:56.217 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:44:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:56 vm08.local ceph-mon[47734]: pgmap v1549: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:56 vm10.local ceph-mon[51861]: pgmap v1549: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:44:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:44:58 vm08.local ceph-mon[47734]: pgmap v1550: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:44:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:44:58 vm10.local ceph-mon[51861]: pgmap v1550: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:01.218 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:01.218 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:01.243 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:01.244 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:00 vm08.local ceph-mon[47734]: pgmap v1551: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:00 vm10.local ceph-mon[51861]: pgmap v1551: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:02 vm08.local ceph-mon[47734]: pgmap v1552: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:02 vm10.local ceph-mon[51861]: pgmap v1552: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:04.966 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:04 vm08.local ceph-mon[47734]: pgmap v1553: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:04.966 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:45:04.966 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:45:04.966 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:45:04.966 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:45:04.966 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:04 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:45:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:04 vm10.local ceph-mon[51861]: pgmap v1553: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:45:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:45:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:45:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:45:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:04 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:45:06.245 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:06.245 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:06.270 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:06.271 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:06.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:06 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:45:06.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:06 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:45:06.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:06 vm10.local ceph-mon[51861]: pgmap v1554: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:06 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:45:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:06 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:45:06.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:06 vm08.local ceph-mon[47734]: pgmap v1554: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:08 vm08.local ceph-mon[47734]: pgmap v1555: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:08 vm10.local ceph-mon[51861]: pgmap v1555: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:11.272 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:11.273 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:11.299 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:11.299 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:10 vm08.local ceph-mon[47734]: pgmap v1556: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:10 vm10.local ceph-mon[51861]: pgmap v1556: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:12 vm08.local ceph-mon[47734]: pgmap v1557: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:12 vm10.local ceph-mon[51861]: pgmap v1557: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:14 vm08.local ceph-mon[47734]: pgmap v1558: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:14 vm10.local ceph-mon[51861]: pgmap v1558: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:16.301 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:16.301 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:16.327 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:16.327 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:16 vm08.local ceph-mon[47734]: pgmap v1559: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:16 vm10.local ceph-mon[51861]: pgmap v1559: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:18 vm08.local ceph-mon[47734]: pgmap v1560: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:18 vm10.local ceph-mon[51861]: pgmap v1560: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:20.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:20 vm08.local ceph-mon[47734]: pgmap v1561: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:20.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:20 vm10.local ceph-mon[51861]: pgmap v1561: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:21.329 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:21.329 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:21.354 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:21.355 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:22 vm08.local ceph-mon[47734]: pgmap v1562: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:22 vm10.local ceph-mon[51861]: pgmap v1562: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:24 vm08.local ceph-mon[47734]: pgmap v1563: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:24 vm10.local ceph-mon[51861]: pgmap v1563: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:26.356 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:26.356 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:26.384 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:26.384 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:26 vm08.local ceph-mon[47734]: pgmap v1564: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:26 vm10.local ceph-mon[51861]: pgmap v1564: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:28 vm08.local ceph-mon[47734]: pgmap v1565: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:28 vm10.local ceph-mon[51861]: pgmap v1565: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:30 vm08.local ceph-mon[47734]: pgmap v1566: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:31.385 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:31.386 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:30 vm10.local ceph-mon[51861]: pgmap v1566: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:31.412 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:31.412 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:32 vm08.local ceph-mon[47734]: pgmap v1567: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:32 vm10.local ceph-mon[51861]: pgmap v1567: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:34.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:34 vm08.local ceph-mon[47734]: pgmap v1568: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:34.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:34 vm10.local ceph-mon[51861]: pgmap v1568: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:36.413 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:36.414 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:36.440 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:36.440 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:36 vm08.local ceph-mon[47734]: pgmap v1569: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:36 vm10.local ceph-mon[51861]: pgmap v1569: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:38 vm08.local ceph-mon[47734]: pgmap v1570: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:38 vm10.local ceph-mon[51861]: pgmap v1570: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:41 vm08.local ceph-mon[47734]: pgmap v1571: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:41 vm10.local ceph-mon[51861]: pgmap v1571: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:41.441 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:41.442 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:41.469 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:41.470 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:42 vm08.local ceph-mon[47734]: pgmap v1572: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:42 vm10.local ceph-mon[51861]: pgmap v1572: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:45:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:45:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:45:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:45:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:44 vm08.local ceph-mon[47734]: pgmap v1573: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:44 vm10.local ceph-mon[51861]: pgmap v1573: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:46.471 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:46.471 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:46.498 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:46.498 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:46 vm08.local ceph-mon[47734]: pgmap v1574: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:46 vm10.local ceph-mon[51861]: pgmap v1574: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:49.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:48 vm08.local ceph-mon[47734]: pgmap v1575: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:48 vm10.local ceph-mon[51861]: pgmap v1575: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:50.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:50 vm10.local ceph-mon[51861]: pgmap v1576: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:50 vm08.local ceph-mon[47734]: pgmap v1576: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:51.500 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:51.500 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:51.602 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:51.602 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:52 vm08.local ceph-mon[47734]: pgmap v1577: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:52 vm10.local ceph-mon[51861]: pgmap v1577: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:54 vm08.local ceph-mon[47734]: pgmap v1578: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:54 vm10.local ceph-mon[51861]: pgmap v1578: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:56.603 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:45:56.604 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:45:56.630 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:45:56.630 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:45:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:56 vm08.local ceph-mon[47734]: pgmap v1579: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:56 vm10.local ceph-mon[51861]: pgmap v1579: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:45:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:45:58 vm08.local ceph-mon[47734]: pgmap v1580: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:45:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:45:58 vm10.local ceph-mon[51861]: pgmap v1580: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:00.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:00 vm08.local ceph-mon[47734]: pgmap v1581: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:00.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:00 vm10.local ceph-mon[51861]: pgmap v1581: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:01.631 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:01.632 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:01.658 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:01.658 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:02 vm08.local ceph-mon[47734]: pgmap v1582: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:02 vm10.local ceph-mon[51861]: pgmap v1582: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:04.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:03 vm08.local ceph-mon[47734]: pgmap v1583: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:04.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:03 vm10.local ceph-mon[51861]: pgmap v1583: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:05 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:46:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:05 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:46:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:05 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:46:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:05 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:46:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:05 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:46:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:05 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:46:06.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:06 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:46:06.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:06 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:46:06.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:06 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:46:06.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:06 vm10.local ceph-mon[51861]: pgmap v1584: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:06.458 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:06 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:46:06.458 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:06 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:46:06.458 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:06 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:46:06.458 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:06 vm08.local ceph-mon[47734]: pgmap v1584: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:06.660 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:06.660 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:06.687 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:06.688 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:08 vm08.local ceph-mon[47734]: pgmap v1585: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:08 vm10.local ceph-mon[51861]: pgmap v1585: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:10.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:10 vm10.local ceph-mon[51861]: pgmap v1586: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:10 vm08.local ceph-mon[47734]: pgmap v1586: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:11.689 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:11.690 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:11.717 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:11.718 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:12 vm08.local ceph-mon[47734]: pgmap v1587: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:12 vm10.local ceph-mon[51861]: pgmap v1587: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:14 vm08.local ceph-mon[47734]: pgmap v1588: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:14 vm10.local ceph-mon[51861]: pgmap v1588: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:16.719 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:16.720 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:16.746 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:16.747 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:16 vm08.local ceph-mon[47734]: pgmap v1589: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:16 vm10.local ceph-mon[51861]: pgmap v1589: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:18 vm08.local ceph-mon[47734]: pgmap v1590: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:18 vm10.local ceph-mon[51861]: pgmap v1590: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:21.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:20 vm08.local ceph-mon[47734]: pgmap v1591: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:20 vm10.local ceph-mon[51861]: pgmap v1591: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:21.748 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:21.749 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:21.774 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:21.775 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:22 vm08.local ceph-mon[47734]: pgmap v1592: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:22 vm10.local ceph-mon[51861]: pgmap v1592: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:24 vm08.local ceph-mon[47734]: pgmap v1593: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:24 vm10.local ceph-mon[51861]: pgmap v1593: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:26.776 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:26.777 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:26.802 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:26.803 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:26 vm08.local ceph-mon[47734]: pgmap v1594: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:26 vm10.local ceph-mon[51861]: pgmap v1594: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:28.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:27 vm08.local ceph-mon[47734]: pgmap v1595: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:28.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:27 vm10.local ceph-mon[51861]: pgmap v1595: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:30 vm10.local ceph-mon[51861]: pgmap v1596: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:30.987 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:30 vm08.local ceph-mon[47734]: pgmap v1596: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:31.804 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:31.805 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:31.831 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:31.832 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:32 vm08.local ceph-mon[47734]: pgmap v1597: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:32 vm10.local ceph-mon[51861]: pgmap v1597: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:34 vm08.local ceph-mon[47734]: pgmap v1598: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:34 vm10.local ceph-mon[51861]: pgmap v1598: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:36.833 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:36.834 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:36.871 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:36.871 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:37.068 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:36 vm08.local ceph-mon[47734]: pgmap v1599: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:36 vm10.local ceph-mon[51861]: pgmap v1599: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:39 vm08.local ceph-mon[47734]: pgmap v1600: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:39 vm10.local ceph-mon[51861]: pgmap v1600: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:40.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:40 vm08.local ceph-mon[47734]: pgmap v1601: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:40.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:40 vm10.local ceph-mon[51861]: pgmap v1601: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:41.873 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:41.873 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:41.901 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:41.901 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:42 vm10.local ceph-mon[51861]: pgmap v1602: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:42 vm08.local ceph-mon[47734]: pgmap v1602: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:46:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:46:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:46:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:46:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:44 vm08.local ceph-mon[47734]: pgmap v1603: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:44 vm10.local ceph-mon[51861]: pgmap v1603: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:46.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:45 vm08.local ceph-mon[47734]: pgmap v1604: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:46.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:45 vm10.local ceph-mon[51861]: pgmap v1604: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:46.903 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:46.903 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:46.929 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:46.930 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:48 vm08.local ceph-mon[47734]: pgmap v1605: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:48 vm10.local ceph-mon[51861]: pgmap v1605: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:50 vm08.local ceph-mon[47734]: pgmap v1606: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:50 vm10.local ceph-mon[51861]: pgmap v1606: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:51.931 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:51.932 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:51.957 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:51.957 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:52 vm08.local ceph-mon[47734]: pgmap v1607: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:52 vm10.local ceph-mon[51861]: pgmap v1607: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:54 vm08.local ceph-mon[47734]: pgmap v1608: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:54 vm10.local ceph-mon[51861]: pgmap v1608: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:56.959 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:46:56.959 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:46:56.986 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:46:56.986 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:46:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:56 vm08.local ceph-mon[47734]: pgmap v1609: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:56 vm10.local ceph-mon[51861]: pgmap v1609: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:46:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:46:57 vm08.local ceph-mon[47734]: pgmap v1610: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:46:58.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:46:57 vm10.local ceph-mon[51861]: pgmap v1610: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:00 vm10.local ceph-mon[51861]: pgmap v1611: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:00 vm08.local ceph-mon[47734]: pgmap v1611: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:01.987 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:01.988 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:02.014 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:02.014 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:02 vm08.local ceph-mon[47734]: pgmap v1612: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:02 vm10.local ceph-mon[51861]: pgmap v1612: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:04 vm08.local ceph-mon[47734]: pgmap v1613: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:04 vm10.local ceph-mon[51861]: pgmap v1613: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:05.617 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:05 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:47:05.617 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:05 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:47:05.617 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:05 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:47:05.640 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:05 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:47:05.640 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:05 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:47:05.640 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:05 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:47:06.438 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:06 vm10.local ceph-mon[51861]: pgmap v1614: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:06.438 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:06 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:47:06.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:06 vm08.local ceph-mon[47734]: pgmap v1614: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:06.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:06 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:47:07.016 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:07.016 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:07.055 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:07.056 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:07 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:47:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:07 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:47:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:07 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:47:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:07 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:47:07.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:07 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:47:07.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:07 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:47:07.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:07 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:47:07.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:07 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:47:08.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:08 vm10.local ceph-mon[51861]: pgmap v1615: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:08.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:08 vm08.local ceph-mon[47734]: pgmap v1615: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:10 vm10.local ceph-mon[51861]: pgmap v1616: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:10 vm08.local ceph-mon[47734]: pgmap v1616: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:12.058 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:12.058 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:12.085 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:12.086 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:13.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:12 vm08.local ceph-mon[47734]: pgmap v1617: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:12 vm10.local ceph-mon[51861]: pgmap v1617: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:14.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:14 vm08.local ceph-mon[47734]: pgmap v1618: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:14.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:14 vm10.local ceph-mon[51861]: pgmap v1618: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:16.792 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:16 vm08.local ceph-mon[47734]: pgmap v1619: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:16 vm10.local ceph-mon[51861]: pgmap v1619: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:17.087 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:17.088 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:17.114 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:17.115 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:19.308 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:18 vm08.local ceph-mon[47734]: pgmap v1620: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:18 vm10.local ceph-mon[51861]: pgmap v1620: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:20 vm08.local ceph-mon[47734]: pgmap v1621: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:20 vm10.local ceph-mon[51861]: pgmap v1621: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:22.116 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:22.117 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:22.142 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:22.143 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:22 vm08.local ceph-mon[47734]: pgmap v1622: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:22 vm10.local ceph-mon[51861]: pgmap v1622: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:24 vm08.local ceph-mon[47734]: pgmap v1623: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:24 vm10.local ceph-mon[51861]: pgmap v1623: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:26.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:25 vm08.local ceph-mon[47734]: pgmap v1624: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:26.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:25 vm10.local ceph-mon[51861]: pgmap v1624: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:27.144 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:27.145 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:27.172 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:27.172 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:28 vm08.local ceph-mon[47734]: pgmap v1625: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:28 vm10.local ceph-mon[51861]: pgmap v1625: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:30 vm08.local ceph-mon[47734]: pgmap v1626: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:30 vm10.local ceph-mon[51861]: pgmap v1626: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:32.174 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:32.174 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:32.201 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:32.201 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:32 vm08.local ceph-mon[47734]: pgmap v1627: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:32 vm10.local ceph-mon[51861]: pgmap v1627: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:34 vm08.local ceph-mon[47734]: pgmap v1628: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:34 vm10.local ceph-mon[51861]: pgmap v1628: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:36.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:36 vm08.local ceph-mon[47734]: pgmap v1629: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:36.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:36 vm10.local ceph-mon[51861]: pgmap v1629: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:37.203 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:37.204 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:37.237 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:37.237 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:39.262 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:38 vm10.local ceph-mon[51861]: pgmap v1630: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:38 vm08.local ceph-mon[47734]: pgmap v1630: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:40.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:40 vm08.local ceph-mon[47734]: pgmap v1631: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:40.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:40 vm10.local ceph-mon[51861]: pgmap v1631: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:42.239 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:42.239 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:42.265 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:42.265 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:42 vm08.local ceph-mon[47734]: pgmap v1632: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:42 vm10.local ceph-mon[51861]: pgmap v1632: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:47:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:47:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:47:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:47:45.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:45 vm08.local ceph-mon[47734]: pgmap v1633: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:45 vm10.local ceph-mon[51861]: pgmap v1633: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:46.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:46 vm08.local ceph-mon[47734]: pgmap v1634: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:46.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:46 vm10.local ceph-mon[51861]: pgmap v1634: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:47.267 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:47.268 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:47.294 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:47.295 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:48.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:48 vm08.local ceph-mon[47734]: pgmap v1635: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:48.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:48 vm10.local ceph-mon[51861]: pgmap v1635: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:50 vm10.local ceph-mon[51861]: pgmap v1636: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:50 vm08.local ceph-mon[47734]: pgmap v1636: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:52.296 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:52.297 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:52.324 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:52.324 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:52 vm08.local ceph-mon[47734]: pgmap v1637: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:52 vm10.local ceph-mon[51861]: pgmap v1637: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:54.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:54 vm08.local ceph-mon[47734]: pgmap v1638: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:54.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:54 vm10.local ceph-mon[51861]: pgmap v1638: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:56 vm08.local ceph-mon[47734]: pgmap v1639: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:56 vm10.local ceph-mon[51861]: pgmap v1639: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:47:57.326 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:47:57.326 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:47:57.372 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:47:57.373 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:47:58.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:47:58 vm08.local ceph-mon[47734]: pgmap v1640: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:47:58.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:47:58 vm10.local ceph-mon[51861]: pgmap v1640: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:00 vm10.local ceph-mon[51861]: pgmap v1641: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:00 vm08.local ceph-mon[47734]: pgmap v1641: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:02.375 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:02.376 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:02.401 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:02.402 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:02 vm10.local ceph-mon[51861]: pgmap v1642: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:02 vm08.local ceph-mon[47734]: pgmap v1642: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:04 vm10.local ceph-mon[51861]: pgmap v1643: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:04 vm08.local ceph-mon[47734]: pgmap v1643: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:06 vm08.local ceph-mon[47734]: pgmap v1644: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:06 vm10.local ceph-mon[51861]: pgmap v1644: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:07.403 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:07.404 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:07.435 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:07.436 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:08.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:48:08.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:48:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:48:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:48:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:48:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:08 vm08.local ceph-mon[47734]: pgmap v1645: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:48:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:48:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:48:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:48:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:48:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:08 vm10.local ceph-mon[51861]: pgmap v1645: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:10 vm10.local ceph-mon[51861]: pgmap v1646: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:10 vm08.local ceph-mon[47734]: pgmap v1646: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:12.437 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:12.438 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:12.464 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:12.464 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:12 vm10.local ceph-mon[51861]: pgmap v1647: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:12 vm08.local ceph-mon[47734]: pgmap v1647: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:14 vm10.local ceph-mon[51861]: pgmap v1648: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:14 vm08.local ceph-mon[47734]: pgmap v1648: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:16 vm08.local ceph-mon[47734]: pgmap v1649: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:16 vm10.local ceph-mon[51861]: pgmap v1649: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:17.466 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:17.466 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:17.491 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:17.492 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:18.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:18 vm08.local ceph-mon[47734]: pgmap v1650: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:18.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:18 vm10.local ceph-mon[51861]: pgmap v1650: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:20 vm10.local ceph-mon[51861]: pgmap v1651: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:20 vm08.local ceph-mon[47734]: pgmap v1651: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:22.493 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:22.494 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:22.519 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:22.519 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:22 vm10.local ceph-mon[51861]: pgmap v1652: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:22 vm08.local ceph-mon[47734]: pgmap v1652: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:24 vm10.local ceph-mon[51861]: pgmap v1653: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:24 vm08.local ceph-mon[47734]: pgmap v1653: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:26 vm08.local ceph-mon[47734]: pgmap v1654: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:26 vm10.local ceph-mon[51861]: pgmap v1654: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:27.520 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:27.521 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:27.547 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:27.548 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:28.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:28 vm08.local ceph-mon[47734]: pgmap v1655: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:28.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:28 vm10.local ceph-mon[51861]: pgmap v1655: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:30 vm10.local ceph-mon[51861]: pgmap v1656: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:30 vm08.local ceph-mon[47734]: pgmap v1656: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:32.550 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:32.550 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:32.576 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:32.577 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:32 vm10.local ceph-mon[51861]: pgmap v1657: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:32 vm08.local ceph-mon[47734]: pgmap v1657: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:34 vm10.local ceph-mon[51861]: pgmap v1658: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:35.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:34 vm08.local ceph-mon[47734]: pgmap v1658: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:36 vm08.local ceph-mon[47734]: pgmap v1659: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:36 vm10.local ceph-mon[51861]: pgmap v1659: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:37.578 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:37.579 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:37.696 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:37.697 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:38.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:38 vm10.local ceph-mon[51861]: pgmap v1660: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:38 vm08.local ceph-mon[47734]: pgmap v1660: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:40 vm10.local ceph-mon[51861]: pgmap v1661: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:40 vm08.local ceph-mon[47734]: pgmap v1661: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:42.698 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:42.699 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:42.724 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:42.724 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:42 vm10.local ceph-mon[51861]: pgmap v1662: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:42 vm08.local ceph-mon[47734]: pgmap v1662: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:48:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:48:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:48:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:48:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:44 vm10.local ceph-mon[51861]: pgmap v1663: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:44 vm08.local ceph-mon[47734]: pgmap v1663: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:46 vm08.local ceph-mon[47734]: pgmap v1664: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:46 vm10.local ceph-mon[51861]: pgmap v1664: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:47.726 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:47.726 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:47.785 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:47.785 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:48.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:48 vm10.local ceph-mon[51861]: pgmap v1665: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:48 vm08.local ceph-mon[47734]: pgmap v1665: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:50 vm10.local ceph-mon[51861]: pgmap v1666: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:50 vm08.local ceph-mon[47734]: pgmap v1666: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:52.787 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:52.787 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:52.814 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:52.814 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:52 vm10.local ceph-mon[51861]: pgmap v1667: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:52 vm08.local ceph-mon[47734]: pgmap v1667: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:54 vm10.local ceph-mon[51861]: pgmap v1668: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:54 vm08.local ceph-mon[47734]: pgmap v1668: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:56 vm08.local ceph-mon[47734]: pgmap v1669: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:56 vm10.local ceph-mon[51861]: pgmap v1669: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:48:57.816 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:48:57.816 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:48:57.843 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:48:57.843 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:48:58.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:48:58 vm08.local ceph-mon[47734]: pgmap v1670: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:48:58.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:48:58 vm10.local ceph-mon[51861]: pgmap v1670: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:00 vm10.local ceph-mon[51861]: pgmap v1671: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:00 vm08.local ceph-mon[47734]: pgmap v1671: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:02.845 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:02.845 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:02.871 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:02.871 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:02 vm10.local ceph-mon[51861]: pgmap v1672: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:02 vm08.local ceph-mon[47734]: pgmap v1672: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:04 vm10.local ceph-mon[51861]: pgmap v1673: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:04 vm08.local ceph-mon[47734]: pgmap v1673: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:06.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:06 vm08.local ceph-mon[47734]: pgmap v1674: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:06 vm10.local ceph-mon[51861]: pgmap v1674: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:07.873 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:07.874 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:07.901 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:07.902 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:49:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:49:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:49:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:08 vm08.local ceph-mon[47734]: pgmap v1675: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:49:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:49:08.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:49:08.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:49:08.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:49:08.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:08 vm10.local ceph-mon[51861]: pgmap v1675: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:08.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:49:08.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:49:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:10 vm10.local ceph-mon[51861]: pgmap v1676: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:10 vm08.local ceph-mon[47734]: pgmap v1676: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:12.904 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:12.904 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:12 vm10.local ceph-mon[51861]: pgmap v1677: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:12.931 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:12.931 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:12 vm08.local ceph-mon[47734]: pgmap v1677: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:14 vm10.local ceph-mon[51861]: pgmap v1678: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:14 vm08.local ceph-mon[47734]: pgmap v1678: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:16 vm08.local ceph-mon[47734]: pgmap v1679: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:16 vm10.local ceph-mon[51861]: pgmap v1679: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:17.933 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:17.934 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:17.960 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:17.961 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:18.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:18 vm08.local ceph-mon[47734]: pgmap v1680: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:18.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:18 vm10.local ceph-mon[51861]: pgmap v1680: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:20 vm10.local ceph-mon[51861]: pgmap v1681: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:20 vm08.local ceph-mon[47734]: pgmap v1681: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:22 vm10.local ceph-mon[51861]: pgmap v1682: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:22.962 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:22.963 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:22.989 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:22.989 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:23.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:22 vm08.local ceph-mon[47734]: pgmap v1682: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:24 vm10.local ceph-mon[51861]: pgmap v1683: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:24 vm08.local ceph-mon[47734]: pgmap v1683: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:26 vm08.local ceph-mon[47734]: pgmap v1684: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:26 vm10.local ceph-mon[51861]: pgmap v1684: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:27.991 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:27.991 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:28.017 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:28.017 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:28.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:28 vm10.local ceph-mon[51861]: pgmap v1685: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:28 vm08.local ceph-mon[47734]: pgmap v1685: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:30 vm10.local ceph-mon[51861]: pgmap v1686: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:30 vm08.local ceph-mon[47734]: pgmap v1686: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:32 vm10.local ceph-mon[51861]: pgmap v1687: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:33.019 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:33.020 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:33.046 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:33.046 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:32 vm08.local ceph-mon[47734]: pgmap v1687: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:34 vm10.local ceph-mon[51861]: pgmap v1688: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:34 vm08.local ceph-mon[47734]: pgmap v1688: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:36 vm08.local ceph-mon[47734]: pgmap v1689: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:36 vm10.local ceph-mon[51861]: pgmap v1689: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:38.048 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:38.048 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:38.073 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:38.074 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:38.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:38 vm10.local ceph-mon[51861]: pgmap v1690: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:38.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:38 vm08.local ceph-mon[47734]: pgmap v1690: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:40 vm10.local ceph-mon[51861]: pgmap v1691: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:40 vm08.local ceph-mon[47734]: pgmap v1691: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:42 vm10.local ceph-mon[51861]: pgmap v1692: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:42 vm08.local ceph-mon[47734]: pgmap v1692: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:43.075 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:43.075 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:43.105 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:43.105 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:49:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:49:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:49:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:49:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:44 vm10.local ceph-mon[51861]: pgmap v1693: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:44 vm08.local ceph-mon[47734]: pgmap v1693: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:46.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:46 vm08.local ceph-mon[47734]: pgmap v1694: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:46 vm10.local ceph-mon[51861]: pgmap v1694: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:48.107 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:48.107 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:48.132 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:48.133 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:48.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:48 vm10.local ceph-mon[51861]: pgmap v1695: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:48.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:48 vm08.local ceph-mon[47734]: pgmap v1695: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:50 vm10.local ceph-mon[51861]: pgmap v1696: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:50 vm08.local ceph-mon[47734]: pgmap v1696: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:52 vm10.local ceph-mon[51861]: pgmap v1697: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:52 vm08.local ceph-mon[47734]: pgmap v1697: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:53.134 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:53.135 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:53.162 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:53.163 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:54 vm10.local ceph-mon[51861]: pgmap v1698: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:54 vm08.local ceph-mon[47734]: pgmap v1698: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:57.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:56 vm08.local ceph-mon[47734]: pgmap v1699: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:56 vm10.local ceph-mon[51861]: pgmap v1699: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:49:58.165 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:49:58.166 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:49:58.192 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:49:58.192 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:49:58.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:49:58 vm10.local ceph-mon[51861]: pgmap v1700: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:49:58.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:49:58 vm08.local ceph-mon[47734]: pgmap v1700: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:00 vm08.local ceph-mon[47734]: pgmap v1701: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T14:50:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:00 vm10.local ceph-mon[51861]: pgmap v1701: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T14:50:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:02 vm08.local ceph-mon[47734]: pgmap v1702: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:02 vm10.local ceph-mon[51861]: pgmap v1702: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:03.194 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:03.195 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:03.223 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:03.224 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:04 vm08.local ceph-mon[47734]: pgmap v1703: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:04 vm10.local ceph-mon[51861]: pgmap v1703: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:07.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:06 vm08.local ceph-mon[47734]: pgmap v1704: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:06 vm10.local ceph-mon[51861]: pgmap v1704: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:08.225 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:08.226 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:08.253 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:08.253 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:08 vm08.local ceph-mon[47734]: pgmap v1705: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:50:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:50:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:50:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:50:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:50:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:08 vm10.local ceph-mon[51861]: pgmap v1705: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:50:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:50:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:50:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:50:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:50:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:10 vm10.local ceph-mon[51861]: pgmap v1706: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:10 vm08.local ceph-mon[47734]: pgmap v1706: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:12 vm10.local ceph-mon[51861]: pgmap v1707: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:12 vm08.local ceph-mon[47734]: pgmap v1707: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:13.255 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:13.256 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:13.282 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:13.283 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:14 vm10.local ceph-mon[51861]: pgmap v1708: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:14 vm08.local ceph-mon[47734]: pgmap v1708: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:16 vm08.local ceph-mon[47734]: pgmap v1709: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:16 vm10.local ceph-mon[51861]: pgmap v1709: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:18.285 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:18.285 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:18.349 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:18.350 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:18.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:18 vm10.local ceph-mon[51861]: pgmap v1710: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:18 vm08.local ceph-mon[47734]: pgmap v1710: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:20 vm10.local ceph-mon[51861]: pgmap v1711: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:20 vm08.local ceph-mon[47734]: pgmap v1711: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:22 vm10.local ceph-mon[51861]: pgmap v1712: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:22 vm08.local ceph-mon[47734]: pgmap v1712: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:23.351 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:23.352 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:23.379 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:23.379 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:24 vm10.local ceph-mon[51861]: pgmap v1713: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:24 vm08.local ceph-mon[47734]: pgmap v1713: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:26.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:26 vm08.local ceph-mon[47734]: pgmap v1714: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:26 vm10.local ceph-mon[51861]: pgmap v1714: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:28.381 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:28.381 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:28.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:28 vm10.local ceph-mon[51861]: pgmap v1715: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:28.408 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:28.409 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:28 vm08.local ceph-mon[47734]: pgmap v1715: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:30 vm10.local ceph-mon[51861]: pgmap v1716: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:30 vm08.local ceph-mon[47734]: pgmap v1716: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:32 vm10.local ceph-mon[51861]: pgmap v1717: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:32 vm08.local ceph-mon[47734]: pgmap v1717: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:33.410 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:33.411 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:33.439 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:33.439 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:34 vm10.local ceph-mon[51861]: pgmap v1718: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:34 vm08.local ceph-mon[47734]: pgmap v1718: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:36 vm08.local ceph-mon[47734]: pgmap v1719: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:36.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:36 vm10.local ceph-mon[51861]: pgmap v1719: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:38.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:38 vm10.local ceph-mon[51861]: pgmap v1720: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:38.440 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:38.441 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:38.468 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:38.469 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:38.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:38 vm08.local ceph-mon[47734]: pgmap v1720: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:40 vm10.local ceph-mon[51861]: pgmap v1721: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:40 vm08.local ceph-mon[47734]: pgmap v1721: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:42 vm10.local ceph-mon[51861]: pgmap v1722: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:42 vm08.local ceph-mon[47734]: pgmap v1722: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:43.470 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:43.471 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:43.498 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:43.498 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:50:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:50:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:50:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:50:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:44 vm08.local ceph-mon[47734]: pgmap v1723: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:44 vm10.local ceph-mon[51861]: pgmap v1723: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:46 vm08.local ceph-mon[47734]: pgmap v1724: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:46 vm10.local ceph-mon[51861]: pgmap v1724: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:48.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:48 vm10.local ceph-mon[51861]: pgmap v1725: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:48.499 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:48.500 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:48.526 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:48.526 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:48 vm08.local ceph-mon[47734]: pgmap v1725: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:50 vm10.local ceph-mon[51861]: pgmap v1726: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:50 vm08.local ceph-mon[47734]: pgmap v1726: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:52 vm10.local ceph-mon[51861]: pgmap v1727: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:52 vm08.local ceph-mon[47734]: pgmap v1727: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:53.527 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:53.528 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:53.555 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:53.556 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:50:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:54 vm10.local ceph-mon[51861]: pgmap v1728: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:54 vm08.local ceph-mon[47734]: pgmap v1728: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:56.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:56 vm08.local ceph-mon[47734]: pgmap v1729: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:56.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:56 vm10.local ceph-mon[51861]: pgmap v1729: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:50:58.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:50:58 vm10.local ceph-mon[51861]: pgmap v1730: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:50:58 vm08.local ceph-mon[47734]: pgmap v1730: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:50:58.557 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:50:58.558 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:50:58.762 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:50:58.763 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:00 vm10.local ceph-mon[51861]: pgmap v1731: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:00 vm08.local ceph-mon[47734]: pgmap v1731: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:02 vm10.local ceph-mon[51861]: pgmap v1732: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:02 vm08.local ceph-mon[47734]: pgmap v1732: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:03.764 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:03.765 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:03.792 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:03.793 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:04 vm10.local ceph-mon[51861]: pgmap v1733: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:04 vm08.local ceph-mon[47734]: pgmap v1733: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:06 vm08.local ceph-mon[47734]: pgmap v1734: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:06 vm10.local ceph-mon[51861]: pgmap v1734: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:08.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:08 vm10.local ceph-mon[51861]: pgmap v1735: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:51:08.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:51:08.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:51:08.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:51:08.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:51:08.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:51:08.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:08 vm08.local ceph-mon[47734]: pgmap v1735: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:51:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:51:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:51:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:51:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:51:08.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:51:08.794 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:08.794 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:08.821 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:08.821 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:10 vm10.local ceph-mon[51861]: pgmap v1736: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:10 vm08.local ceph-mon[47734]: pgmap v1736: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:12 vm10.local ceph-mon[51861]: pgmap v1737: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:51:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:12 vm08.local ceph-mon[47734]: pgmap v1737: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:51:13.822 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:13.823 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:13.849 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:13.850 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:14 vm10.local ceph-mon[51861]: pgmap v1738: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:51:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:14 vm08.local ceph-mon[47734]: pgmap v1738: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T14:51:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:16 vm08.local ceph-mon[47734]: pgmap v1739: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:16 vm10.local ceph-mon[51861]: pgmap v1739: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:18.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:18 vm10.local ceph-mon[51861]: pgmap v1740: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:18 vm08.local ceph-mon[47734]: pgmap v1740: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:18.851 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:18.852 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:18.878 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:18.879 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:20 vm10.local ceph-mon[51861]: pgmap v1741: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:20 vm08.local ceph-mon[47734]: pgmap v1741: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:22 vm10.local ceph-mon[51861]: pgmap v1742: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:22 vm08.local ceph-mon[47734]: pgmap v1742: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:23.880 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:23.880 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:23.907 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:23.908 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:24 vm10.local ceph-mon[51861]: pgmap v1743: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:24 vm08.local ceph-mon[47734]: pgmap v1743: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:26.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:26 vm08.local ceph-mon[47734]: pgmap v1744: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:26.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:26 vm10.local ceph-mon[51861]: pgmap v1744: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:28 vm08.local ceph-mon[47734]: pgmap v1745: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:28 vm10.local ceph-mon[51861]: pgmap v1745: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:28.909 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:28.909 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:28.935 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:28.935 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:30 vm10.local ceph-mon[51861]: pgmap v1746: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:31.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:30 vm08.local ceph-mon[47734]: pgmap v1746: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:32.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:32 vm10.local ceph-mon[51861]: pgmap v1747: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:32 vm08.local ceph-mon[47734]: pgmap v1747: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:33.936 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:33.937 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:33.965 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:33.966 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:34 vm10.local ceph-mon[51861]: pgmap v1748: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:34 vm08.local ceph-mon[47734]: pgmap v1748: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:36 vm08.local ceph-mon[47734]: pgmap v1749: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:36 vm10.local ceph-mon[51861]: pgmap v1749: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:38 vm08.local ceph-mon[47734]: pgmap v1750: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:38 vm10.local ceph-mon[51861]: pgmap v1750: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:38.967 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:38.993 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:39.164 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:39.165 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:40 vm10.local ceph-mon[51861]: pgmap v1751: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:40 vm08.local ceph-mon[47734]: pgmap v1751: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:42 vm10.local ceph-mon[51861]: pgmap v1752: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:42 vm08.local ceph-mon[47734]: pgmap v1752: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:51:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:51:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:51:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:51:44.166 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:44.167 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:44.193 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:44.193 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:44 vm10.local ceph-mon[51861]: pgmap v1753: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:44 vm08.local ceph-mon[47734]: pgmap v1753: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:47.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:46 vm08.local ceph-mon[47734]: pgmap v1754: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:46 vm10.local ceph-mon[51861]: pgmap v1754: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:48 vm08.local ceph-mon[47734]: pgmap v1755: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:48 vm10.local ceph-mon[51861]: pgmap v1755: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:49.195 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:49.195 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:49.223 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:49.224 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:50 vm10.local ceph-mon[51861]: pgmap v1756: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:50 vm08.local ceph-mon[47734]: pgmap v1756: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:52 vm10.local ceph-mon[51861]: pgmap v1757: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:52 vm08.local ceph-mon[47734]: pgmap v1757: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:54.225 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:54.226 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:54.254 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:54.255 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:51:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:54 vm10.local ceph-mon[51861]: pgmap v1758: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:54 vm08.local ceph-mon[47734]: pgmap v1758: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:56 vm08.local ceph-mon[47734]: pgmap v1759: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:56 vm10.local ceph-mon[51861]: pgmap v1759: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:51:58.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:51:58 vm08.local ceph-mon[47734]: pgmap v1760: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:51:58 vm10.local ceph-mon[51861]: pgmap v1760: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:51:59.256 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:51:59.257 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:51:59.285 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:51:59.286 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:00 vm10.local ceph-mon[51861]: pgmap v1761: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:00 vm08.local ceph-mon[47734]: pgmap v1761: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:02 vm10.local ceph-mon[51861]: pgmap v1762: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:02 vm08.local ceph-mon[47734]: pgmap v1762: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:04.287 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:04.288 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:04.314 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:04.314 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:04 vm08.local ceph-mon[47734]: pgmap v1763: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:04 vm10.local ceph-mon[51861]: pgmap v1763: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:06 vm08.local ceph-mon[47734]: pgmap v1764: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:06 vm10.local ceph-mon[51861]: pgmap v1764: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:08 vm10.local ceph-mon[51861]: pgmap v1765: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:52:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:52:08.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:08 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:52:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:08 vm08.local ceph-mon[47734]: pgmap v1765: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:52:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:52:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:08 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:52:09.315 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:09.316 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:09.342 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:09.343 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:52:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:52:09.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:52:09.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:52:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:10 vm10.local ceph-mon[51861]: pgmap v1766: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:10.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:10 vm08.local ceph-mon[47734]: pgmap v1766: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:12 vm10.local ceph-mon[51861]: pgmap v1767: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:12 vm08.local ceph-mon[47734]: pgmap v1767: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:14.345 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:14.345 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:14.472 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:14.473 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:14 vm10.local ceph-mon[51861]: pgmap v1768: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:14 vm08.local ceph-mon[47734]: pgmap v1768: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:16 vm08.local ceph-mon[47734]: pgmap v1769: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:16 vm10.local ceph-mon[51861]: pgmap v1769: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:18 vm08.local ceph-mon[47734]: pgmap v1770: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:18 vm10.local ceph-mon[51861]: pgmap v1770: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:19.474 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:19.475 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:19.501 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:19.501 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:20 vm10.local ceph-mon[51861]: pgmap v1771: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:20 vm08.local ceph-mon[47734]: pgmap v1771: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:22 vm10.local ceph-mon[51861]: pgmap v1772: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:23.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:22 vm08.local ceph-mon[47734]: pgmap v1772: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:24.503 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:24.504 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:24.531 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:24.531 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:24 vm08.local ceph-mon[47734]: pgmap v1773: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:24 vm10.local ceph-mon[51861]: pgmap v1773: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:26 vm08.local ceph-mon[47734]: pgmap v1774: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:26 vm10.local ceph-mon[51861]: pgmap v1774: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:28 vm08.local ceph-mon[47734]: pgmap v1775: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:28 vm10.local ceph-mon[51861]: pgmap v1775: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:29.532 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:29.533 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:29.559 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:29.560 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:30 vm10.local ceph-mon[51861]: pgmap v1776: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:30 vm08.local ceph-mon[47734]: pgmap v1776: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:33.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:32 vm08.local ceph-mon[47734]: pgmap v1777: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:32 vm10.local ceph-mon[51861]: pgmap v1777: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:34.561 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:34.562 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:34.589 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:34.590 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:35.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:34 vm08.local ceph-mon[47734]: pgmap v1778: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:34 vm10.local ceph-mon[51861]: pgmap v1778: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:36 vm08.local ceph-mon[47734]: pgmap v1779: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:36 vm10.local ceph-mon[51861]: pgmap v1779: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:38 vm08.local ceph-mon[47734]: pgmap v1780: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:38 vm10.local ceph-mon[51861]: pgmap v1780: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:39.591 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:39.592 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:39.623 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:39.624 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:41.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:40 vm08.local ceph-mon[47734]: pgmap v1781: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:40 vm10.local ceph-mon[51861]: pgmap v1781: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:42 vm08.local ceph-mon[47734]: pgmap v1782: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:42 vm10.local ceph-mon[51861]: pgmap v1782: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:52:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:52:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:52:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:52:44.625 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:44.626 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:44.652 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:44.652 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:44 vm08.local ceph-mon[47734]: pgmap v1783: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:44 vm10.local ceph-mon[51861]: pgmap v1783: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:46 vm08.local ceph-mon[47734]: pgmap v1784: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:46 vm10.local ceph-mon[51861]: pgmap v1784: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:48 vm08.local ceph-mon[47734]: pgmap v1785: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:48 vm10.local ceph-mon[51861]: pgmap v1785: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:49.654 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:49.654 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:49.683 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:49.684 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:50 vm08.local ceph-mon[47734]: pgmap v1786: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:50 vm10.local ceph-mon[51861]: pgmap v1786: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:52 vm08.local ceph-mon[47734]: pgmap v1787: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:52 vm10.local ceph-mon[51861]: pgmap v1787: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:54.686 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:54.686 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:54.712 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:54.713 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:52:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:54 vm08.local ceph-mon[47734]: pgmap v1788: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:54 vm10.local ceph-mon[51861]: pgmap v1788: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:56 vm08.local ceph-mon[47734]: pgmap v1789: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:56 vm10.local ceph-mon[51861]: pgmap v1789: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:52:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:52:58 vm08.local ceph-mon[47734]: pgmap v1790: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:52:58 vm10.local ceph-mon[51861]: pgmap v1790: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:52:59.714 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:52:59.714 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:52:59.740 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:52:59.741 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:00 vm08.local ceph-mon[47734]: pgmap v1791: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:00 vm10.local ceph-mon[51861]: pgmap v1791: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:02 vm08.local ceph-mon[47734]: pgmap v1792: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:02 vm10.local ceph-mon[51861]: pgmap v1792: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:04.742 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:04.743 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:04.771 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:04.772 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:04 vm08.local ceph-mon[47734]: pgmap v1793: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:04 vm10.local ceph-mon[51861]: pgmap v1793: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:06.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:06 vm08.local ceph-mon[47734]: pgmap v1794: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:06 vm10.local ceph-mon[51861]: pgmap v1794: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:08.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:08 vm08.local ceph-mon[47734]: pgmap v1795: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:08.625 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:08 vm10.local ceph-mon[51861]: pgmap v1795: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:53:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:53:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:53:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:53:09.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:53:09.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:53:09.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:53:09.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:53:09.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:53:09.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:53:09.773 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:09.774 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:09.800 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:09.800 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:10.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:10 vm08.local ceph-mon[47734]: pgmap v1796: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:10 vm10.local ceph-mon[51861]: pgmap v1796: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:13.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:12 vm08.local ceph-mon[47734]: pgmap v1797: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:12 vm10.local ceph-mon[51861]: pgmap v1797: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:14.802 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:14.803 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:14.829 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:14.829 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:14 vm08.local ceph-mon[47734]: pgmap v1798: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:14 vm10.local ceph-mon[51861]: pgmap v1798: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:16 vm08.local ceph-mon[47734]: pgmap v1799: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:16 vm10.local ceph-mon[51861]: pgmap v1799: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:18.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:18 vm08.local ceph-mon[47734]: pgmap v1800: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:18 vm10.local ceph-mon[51861]: pgmap v1800: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:19.831 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:19.831 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:19.857 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:19.857 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:20 vm08.local ceph-mon[47734]: pgmap v1801: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:20 vm10.local ceph-mon[51861]: pgmap v1801: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:23.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:22 vm08.local ceph-mon[47734]: pgmap v1802: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:22 vm10.local ceph-mon[51861]: pgmap v1802: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:24.859 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:24.859 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:24.886 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:24.887 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:25.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:24 vm08.local ceph-mon[47734]: pgmap v1803: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:24 vm10.local ceph-mon[51861]: pgmap v1803: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:27.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:26 vm08.local ceph-mon[47734]: pgmap v1804: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:26 vm10.local ceph-mon[51861]: pgmap v1804: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:28.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:28 vm08.local ceph-mon[47734]: pgmap v1805: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:28 vm10.local ceph-mon[51861]: pgmap v1805: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:29.888 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:29.888 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:29.914 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:29.914 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:31.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:30 vm08.local ceph-mon[47734]: pgmap v1806: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:30 vm10.local ceph-mon[51861]: pgmap v1806: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:33.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:32 vm08.local ceph-mon[47734]: pgmap v1807: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:32 vm10.local ceph-mon[51861]: pgmap v1807: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:34.916 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:34.916 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:34.944 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:34.944 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:34 vm08.local ceph-mon[47734]: pgmap v1808: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:34 vm10.local ceph-mon[51861]: pgmap v1808: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:36 vm08.local ceph-mon[47734]: pgmap v1809: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:36 vm10.local ceph-mon[51861]: pgmap v1809: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:38.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:38 vm08.local ceph-mon[47734]: pgmap v1810: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:38 vm10.local ceph-mon[51861]: pgmap v1810: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:39.945 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:39.946 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:39.972 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:39.972 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:41.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:40 vm08.local ceph-mon[47734]: pgmap v1811: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:40 vm10.local ceph-mon[51861]: pgmap v1811: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:42 vm08.local ceph-mon[47734]: pgmap v1812: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:42 vm10.local ceph-mon[51861]: pgmap v1812: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:53:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:53:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:53:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:53:44.973 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:44.974 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:45.000 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:45.001 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:44 vm08.local ceph-mon[47734]: pgmap v1813: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:44 vm10.local ceph-mon[51861]: pgmap v1813: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:46 vm08.local ceph-mon[47734]: pgmap v1814: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:46 vm10.local ceph-mon[51861]: pgmap v1814: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:48 vm08.local ceph-mon[47734]: pgmap v1815: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:48 vm10.local ceph-mon[51861]: pgmap v1815: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:50.003 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:50.003 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:50.029 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:50.029 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:50 vm08.local ceph-mon[47734]: pgmap v1816: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:50 vm10.local ceph-mon[51861]: pgmap v1816: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:52 vm08.local ceph-mon[47734]: pgmap v1817: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:52 vm10.local ceph-mon[51861]: pgmap v1817: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:55.031 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:53:55.031 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:53:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:54 vm08.local ceph-mon[47734]: pgmap v1818: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:55.058 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:53:55.058 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:53:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:54 vm10.local ceph-mon[51861]: pgmap v1818: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:56.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:56 vm08.local ceph-mon[47734]: pgmap v1819: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:56 vm10.local ceph-mon[51861]: pgmap v1819: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:53:58.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:53:58 vm08.local ceph-mon[47734]: pgmap v1820: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:53:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:53:58 vm10.local ceph-mon[51861]: pgmap v1820: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:00.060 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:00.060 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:00.087 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:00.088 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:00 vm08.local ceph-mon[47734]: pgmap v1821: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:00 vm10.local ceph-mon[51861]: pgmap v1821: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:02 vm08.local ceph-mon[47734]: pgmap v1822: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:02 vm10.local ceph-mon[51861]: pgmap v1822: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:04 vm08.local ceph-mon[47734]: pgmap v1823: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:05.089 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:05.090 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:05.120 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:05.120 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:04 vm10.local ceph-mon[51861]: pgmap v1823: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:06 vm10.local ceph-mon[51861]: pgmap v1824: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:07.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:06 vm08.local ceph-mon[47734]: pgmap v1824: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:08 vm10.local ceph-mon[51861]: pgmap v1825: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:08.670 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:08 vm08.local ceph-mon[47734]: pgmap v1825: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:09.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:54:09.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:54:09.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:54:09.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:54:09.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:54:09.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:54:09.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:54:09.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:54:09.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:54:09.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:54:10.121 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:10.122 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:10.148 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:10.149 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:10 vm08.local ceph-mon[47734]: pgmap v1826: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:10 vm10.local ceph-mon[51861]: pgmap v1826: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:12 vm08.local ceph-mon[47734]: pgmap v1827: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:12 vm10.local ceph-mon[51861]: pgmap v1827: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:14 vm08.local ceph-mon[47734]: pgmap v1828: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:15.150 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:15.151 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:14 vm10.local ceph-mon[51861]: pgmap v1828: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:15.179 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:15.180 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:17.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:16 vm08.local ceph-mon[47734]: pgmap v1829: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:16 vm10.local ceph-mon[51861]: pgmap v1829: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:18 vm08.local ceph-mon[47734]: pgmap v1830: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:18 vm10.local ceph-mon[51861]: pgmap v1830: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:20.182 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:20.182 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:20.209 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:20.210 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:20 vm08.local ceph-mon[47734]: pgmap v1831: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:20 vm10.local ceph-mon[51861]: pgmap v1831: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:22 vm08.local ceph-mon[47734]: pgmap v1832: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:22 vm10.local ceph-mon[51861]: pgmap v1832: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:24 vm08.local ceph-mon[47734]: pgmap v1833: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:24 vm10.local ceph-mon[51861]: pgmap v1833: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:25.211 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:25.211 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:25.237 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:25.237 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:26 vm08.local ceph-mon[47734]: pgmap v1834: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:26 vm10.local ceph-mon[51861]: pgmap v1834: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:28 vm08.local ceph-mon[47734]: pgmap v1835: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:28 vm10.local ceph-mon[51861]: pgmap v1835: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:30.239 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:30.239 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:30.268 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:30.269 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:31.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:30 vm08.local ceph-mon[47734]: pgmap v1836: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:30 vm10.local ceph-mon[51861]: pgmap v1836: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:32 vm08.local ceph-mon[47734]: pgmap v1837: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:32 vm10.local ceph-mon[51861]: pgmap v1837: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:35.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:34 vm08.local ceph-mon[47734]: pgmap v1838: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:34 vm10.local ceph-mon[51861]: pgmap v1838: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:35.270 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:35.271 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:35.296 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:35.297 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:37.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:36 vm08.local ceph-mon[47734]: pgmap v1839: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:36 vm10.local ceph-mon[51861]: pgmap v1839: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:38 vm08.local ceph-mon[47734]: pgmap v1840: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:38 vm10.local ceph-mon[51861]: pgmap v1840: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:40.298 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:40.298 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:40.325 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:40.325 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:40 vm08.local ceph-mon[47734]: pgmap v1841: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:40 vm10.local ceph-mon[51861]: pgmap v1841: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:42 vm08.local ceph-mon[47734]: pgmap v1842: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:42 vm10.local ceph-mon[51861]: pgmap v1842: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:54:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:54:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:54:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:54:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:44 vm08.local ceph-mon[47734]: pgmap v1843: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:44 vm10.local ceph-mon[51861]: pgmap v1843: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:45.327 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:45.327 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:45.353 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:45.354 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:46 vm08.local ceph-mon[47734]: pgmap v1844: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:46 vm10.local ceph-mon[51861]: pgmap v1844: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:48 vm08.local ceph-mon[47734]: pgmap v1845: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:48 vm10.local ceph-mon[51861]: pgmap v1845: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:50.355 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:50.355 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:50.382 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:50.382 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:50 vm08.local ceph-mon[47734]: pgmap v1846: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:50 vm10.local ceph-mon[51861]: pgmap v1846: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:52 vm08.local ceph-mon[47734]: pgmap v1847: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:52 vm10.local ceph-mon[51861]: pgmap v1847: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:54 vm08.local ceph-mon[47734]: pgmap v1848: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:54 vm10.local ceph-mon[51861]: pgmap v1848: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:55.383 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:54:55.384 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:54:55.410 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:54:55.410 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:54:57.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:56 vm08.local ceph-mon[47734]: pgmap v1849: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:56 vm10.local ceph-mon[51861]: pgmap v1849: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:54:58.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:54:58 vm08.local ceph-mon[47734]: pgmap v1850: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:54:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:54:58 vm10.local ceph-mon[51861]: pgmap v1850: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:00.412 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:00.412 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:00.439 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:00.439 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:00 vm08.local ceph-mon[47734]: pgmap v1851: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:00 vm10.local ceph-mon[51861]: pgmap v1851: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:02 vm08.local ceph-mon[47734]: pgmap v1852: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:02 vm10.local ceph-mon[51861]: pgmap v1852: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:04 vm10.local ceph-mon[51861]: pgmap v1853: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:04 vm08.local ceph-mon[47734]: pgmap v1853: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:05.441 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:05.441 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:05.590 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:05.590 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:06 vm08.local ceph-mon[47734]: pgmap v1854: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:06 vm10.local ceph-mon[51861]: pgmap v1854: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:08.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:08 vm08.local ceph-mon[47734]: pgmap v1855: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:08 vm10.local ceph-mon[51861]: pgmap v1855: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:09.495 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:55:09.495 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:55:09.495 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:55:09.495 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:09.495 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:09.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:55:09.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:55:09.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:55:09.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:09.500 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:10.592 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:10.592 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:10.618 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:10.619 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:10 vm10.local ceph-mon[51861]: pgmap v1856: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:10 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:10 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:10 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:10 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:10 vm08.local ceph-mon[47734]: pgmap v1856: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:10 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:10 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:10 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:10 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:55:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:12 vm08.local ceph-mon[47734]: pgmap v1857: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:12 vm10.local ceph-mon[51861]: pgmap v1857: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:14 vm08.local ceph-mon[47734]: pgmap v1858: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:14 vm10.local ceph-mon[51861]: pgmap v1858: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:15.620 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:15.621 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:15.648 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:15.649 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:16.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:16 vm08.local ceph-mon[47734]: pgmap v1859: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:16 vm10.local ceph-mon[51861]: pgmap v1859: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:18.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:18 vm08.local ceph-mon[47734]: pgmap v1860: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:18 vm10.local ceph-mon[51861]: pgmap v1860: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:20.650 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:20.651 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:20.676 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:20.677 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:20 vm08.local ceph-mon[47734]: pgmap v1861: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:20 vm10.local ceph-mon[51861]: pgmap v1861: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:23.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:22 vm08.local ceph-mon[47734]: pgmap v1862: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:22 vm10.local ceph-mon[51861]: pgmap v1862: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:24 vm08.local ceph-mon[47734]: pgmap v1863: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:24 vm10.local ceph-mon[51861]: pgmap v1863: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:25.678 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:25.679 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:25.707 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:25.708 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:26 vm08.local ceph-mon[47734]: pgmap v1864: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:26 vm10.local ceph-mon[51861]: pgmap v1864: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:28.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:28 vm08.local ceph-mon[47734]: pgmap v1865: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:28 vm10.local ceph-mon[51861]: pgmap v1865: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:30.709 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:30.709 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:30.739 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:30.740 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:30 vm08.local ceph-mon[47734]: pgmap v1866: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:30 vm10.local ceph-mon[51861]: pgmap v1866: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:33.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:32 vm08.local ceph-mon[47734]: pgmap v1867: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:32 vm10.local ceph-mon[51861]: pgmap v1867: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:35.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:34 vm08.local ceph-mon[47734]: pgmap v1868: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:34 vm10.local ceph-mon[51861]: pgmap v1868: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:35.741 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:35.742 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:35.770 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:35.771 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:37.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:36 vm08.local ceph-mon[47734]: pgmap v1869: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:36 vm10.local ceph-mon[51861]: pgmap v1869: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:38.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:38 vm08.local ceph-mon[47734]: pgmap v1870: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:38 vm10.local ceph-mon[51861]: pgmap v1870: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:40.772 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:40.773 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:40.798 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:40.799 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:40 vm08.local ceph-mon[47734]: pgmap v1871: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:40 vm10.local ceph-mon[51861]: pgmap v1871: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:42 vm10.local ceph-mon[51861]: pgmap v1872: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:42 vm08.local ceph-mon[47734]: pgmap v1872: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:55:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:55:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:55:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:55:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:44 vm10.local ceph-mon[51861]: pgmap v1873: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:44 vm08.local ceph-mon[47734]: pgmap v1873: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:45.800 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:45.801 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:45.827 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:45.828 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:46 vm10.local ceph-mon[51861]: pgmap v1874: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:46 vm08.local ceph-mon[47734]: pgmap v1874: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:48.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:48 vm08.local ceph-mon[47734]: pgmap v1875: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:48 vm10.local ceph-mon[51861]: pgmap v1875: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:50.830 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:50.830 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:50.858 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:50.859 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:50 vm08.local ceph-mon[47734]: pgmap v1876: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:50 vm10.local ceph-mon[51861]: pgmap v1876: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:52 vm08.local ceph-mon[47734]: pgmap v1877: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:52 vm10.local ceph-mon[51861]: pgmap v1877: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:54 vm10.local ceph-mon[51861]: pgmap v1878: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:54 vm08.local ceph-mon[47734]: pgmap v1878: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:55.860 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:55:55.861 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:55:55.886 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:55:55.887 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:55:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:56 vm10.local ceph-mon[51861]: pgmap v1879: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:57.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:56 vm08.local ceph-mon[47734]: pgmap v1879: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:55:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:55:58 vm10.local ceph-mon[51861]: pgmap v1880: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:55:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:55:58 vm08.local ceph-mon[47734]: pgmap v1880: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:00.889 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:00.889 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:00.915 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:00.915 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:00 vm08.local ceph-mon[47734]: pgmap v1881: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:00 vm10.local ceph-mon[51861]: pgmap v1881: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:02 vm08.local ceph-mon[47734]: pgmap v1882: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:02 vm10.local ceph-mon[51861]: pgmap v1882: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:04 vm08.local ceph-mon[47734]: pgmap v1883: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:04 vm10.local ceph-mon[51861]: pgmap v1883: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:05.917 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:05.917 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:05.945 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:05.945 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:06 vm08.local ceph-mon[47734]: pgmap v1884: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:06 vm10.local ceph-mon[51861]: pgmap v1884: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:08 vm10.local ceph-mon[51861]: pgmap v1885: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:08.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:08 vm08.local ceph-mon[47734]: pgmap v1885: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:09.962 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:56:09.962 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:56:09.962 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:09 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:56:10.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:56:10.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:56:10.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:09 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:56:10.947 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:10.948 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:10.979 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:10.980 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:10 vm08.local ceph-mon[47734]: pgmap v1886: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:10 vm10.local ceph-mon[51861]: pgmap v1886: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:12.521 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:12 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:56:12.521 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:12 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:56:12.521 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:12 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:56:12.521 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:12 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:56:12.521 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:12 vm08.local ceph-mon[47734]: pgmap v1887: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:12 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:56:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:12 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:56:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:12 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:56:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:12 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:56:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:12 vm10.local ceph-mon[51861]: pgmap v1887: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:14 vm08.local ceph-mon[47734]: pgmap v1888: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:14 vm10.local ceph-mon[51861]: pgmap v1888: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:15.981 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:15.982 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:16.012 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:16.013 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:16 vm08.local ceph-mon[47734]: pgmap v1889: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:16 vm10.local ceph-mon[51861]: pgmap v1889: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:18 vm10.local ceph-mon[51861]: pgmap v1890: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:18 vm08.local ceph-mon[47734]: pgmap v1890: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:21.014 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:21.014 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:21.044 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:21.044 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:20 vm08.local ceph-mon[47734]: pgmap v1891: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:20 vm10.local ceph-mon[51861]: pgmap v1891: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:22.964 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:22 vm08.local ceph-mon[47734]: pgmap v1892: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:22 vm10.local ceph-mon[51861]: pgmap v1892: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:25.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:24 vm08.local ceph-mon[47734]: pgmap v1893: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:24 vm10.local ceph-mon[51861]: pgmap v1893: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:26.046 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:26.047 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:26.073 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:26.074 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:27.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:26 vm08.local ceph-mon[47734]: pgmap v1894: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:26 vm10.local ceph-mon[51861]: pgmap v1894: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:28 vm10.local ceph-mon[51861]: pgmap v1895: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:28.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:28 vm08.local ceph-mon[47734]: pgmap v1895: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:31.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:30 vm08.local ceph-mon[47734]: pgmap v1896: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:31.075 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:31.076 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:31.107 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:31.107 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:30 vm10.local ceph-mon[51861]: pgmap v1896: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:33.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:32 vm08.local ceph-mon[47734]: pgmap v1897: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:32 vm10.local ceph-mon[51861]: pgmap v1897: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:35.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:34 vm08.local ceph-mon[47734]: pgmap v1898: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:34 vm10.local ceph-mon[51861]: pgmap v1898: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:36.109 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:36.109 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:36.137 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:36.137 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:36 vm08.local ceph-mon[47734]: pgmap v1899: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:36 vm10.local ceph-mon[51861]: pgmap v1899: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:38 vm10.local ceph-mon[51861]: pgmap v1900: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:38 vm08.local ceph-mon[47734]: pgmap v1900: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:41.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:40 vm08.local ceph-mon[47734]: pgmap v1901: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:41.139 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:41.139 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:40 vm10.local ceph-mon[51861]: pgmap v1901: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:41.165 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:41.165 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:43.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:42 vm08.local ceph-mon[47734]: pgmap v1902: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:42 vm10.local ceph-mon[51861]: pgmap v1902: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:56:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:56:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:56:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:56:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:44 vm08.local ceph-mon[47734]: pgmap v1903: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:44 vm10.local ceph-mon[51861]: pgmap v1903: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:46.166 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:46.167 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:46.198 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:46.198 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:46.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:46 vm08.local ceph-mon[47734]: pgmap v1904: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:46 vm10.local ceph-mon[51861]: pgmap v1904: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:48 vm10.local ceph-mon[51861]: pgmap v1905: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:48 vm08.local ceph-mon[47734]: pgmap v1905: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:50 vm08.local ceph-mon[47734]: pgmap v1906: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:50 vm10.local ceph-mon[51861]: pgmap v1906: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:51.200 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:51.200 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:51.227 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:51.228 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:52 vm08.local ceph-mon[47734]: pgmap v1907: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:52 vm10.local ceph-mon[51861]: pgmap v1907: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:55.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:54 vm08.local ceph-mon[47734]: pgmap v1908: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:54 vm10.local ceph-mon[51861]: pgmap v1908: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:56.229 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:56:56.230 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:56:56.256 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:56:56.257 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:56:57.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:56 vm08.local ceph-mon[47734]: pgmap v1909: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:56 vm10.local ceph-mon[51861]: pgmap v1909: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:56:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:56:58 vm10.local ceph-mon[51861]: pgmap v1910: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:56:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:56:58 vm08.local ceph-mon[47734]: pgmap v1910: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:00 vm08.local ceph-mon[47734]: pgmap v1911: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:00 vm10.local ceph-mon[51861]: pgmap v1911: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:01.258 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:01.259 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:01.284 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:01.285 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:02 vm08.local ceph-mon[47734]: pgmap v1912: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:02 vm10.local ceph-mon[51861]: pgmap v1912: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:04 vm08.local ceph-mon[47734]: pgmap v1913: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:04 vm10.local ceph-mon[51861]: pgmap v1913: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:06.286 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:06.287 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:06.314 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:06.315 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:07.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:06 vm08.local ceph-mon[47734]: pgmap v1914: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:06 vm10.local ceph-mon[51861]: pgmap v1914: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:08.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:08 vm10.local ceph-mon[51861]: pgmap v1915: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:08.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:08 vm08.local ceph-mon[47734]: pgmap v1915: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:10 vm08.local ceph-mon[47734]: pgmap v1916: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:10 vm10.local ceph-mon[51861]: pgmap v1916: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:11.316 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:11.316 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:11.360 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:11.360 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:11.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:11 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:57:11.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:11 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:57:11.909 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:11 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:57:11.969 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:11 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:57:11.969 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:11 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:57:11.969 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:11 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:57:12.860 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:12 vm10.local ceph-mon[51861]: pgmap v1917: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:12.860 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:12 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:13.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:12 vm08.local ceph-mon[47734]: pgmap v1917: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:13.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:12 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:14.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:13 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:14.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:13 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:15.223 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:14 vm10.local ceph-mon[51861]: pgmap v1918: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:15.223 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:15.223 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:57:15.224 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:15.224 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:15.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:14 vm08.local ceph-mon[47734]: pgmap v1918: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:15.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:15.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T14:57:15.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:15.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:57:16.362 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:16.363 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:16.392 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:16.393 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:17 vm08.local ceph-mon[47734]: pgmap v1919: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:17 vm10.local ceph-mon[51861]: pgmap v1919: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:18.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:18 vm10.local ceph-mon[51861]: pgmap v1920: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:18.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:18 vm08.local ceph-mon[47734]: pgmap v1920: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:20 vm08.local ceph-mon[47734]: pgmap v1921: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:20 vm10.local ceph-mon[51861]: pgmap v1921: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:21.394 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:21.394 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:21.420 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:21.421 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:22 vm08.local ceph-mon[47734]: pgmap v1922: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:22 vm10.local ceph-mon[51861]: pgmap v1922: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:24 vm08.local ceph-mon[47734]: pgmap v1923: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:24 vm10.local ceph-mon[51861]: pgmap v1923: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:26.423 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:26.423 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:26.453 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:26.454 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:27.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:26 vm08.local ceph-mon[47734]: pgmap v1924: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:26 vm10.local ceph-mon[51861]: pgmap v1924: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:28.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:28 vm10.local ceph-mon[51861]: pgmap v1925: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:28.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:28 vm08.local ceph-mon[47734]: pgmap v1925: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:30 vm08.local ceph-mon[47734]: pgmap v1926: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:30 vm10.local ceph-mon[51861]: pgmap v1926: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:31.455 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:31.456 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:31.483 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:31.484 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:33.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:32 vm08.local ceph-mon[47734]: pgmap v1927: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:32 vm10.local ceph-mon[51861]: pgmap v1927: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:34 vm08.local ceph-mon[47734]: pgmap v1928: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:34 vm10.local ceph-mon[51861]: pgmap v1928: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:36.485 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:36.485 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:36.512 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:36.512 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:37.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:36 vm08.local ceph-mon[47734]: pgmap v1929: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:36 vm10.local ceph-mon[51861]: pgmap v1929: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:38 vm10.local ceph-mon[51861]: pgmap v1930: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:38 vm08.local ceph-mon[47734]: pgmap v1930: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:40 vm08.local ceph-mon[47734]: pgmap v1931: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:40 vm10.local ceph-mon[51861]: pgmap v1931: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:41.514 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:41.514 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:41.541 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:41.541 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:42 vm08.local ceph-mon[47734]: pgmap v1932: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:42 vm10.local ceph-mon[51861]: pgmap v1932: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:57:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:57:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:57:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:57:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:44 vm08.local ceph-mon[47734]: pgmap v1933: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:44 vm10.local ceph-mon[51861]: pgmap v1933: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:46.543 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:46.543 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:46.569 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:46.569 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:47.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:46 vm08.local ceph-mon[47734]: pgmap v1934: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:46 vm10.local ceph-mon[51861]: pgmap v1934: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:48.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:48 vm10.local ceph-mon[51861]: pgmap v1935: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:48 vm08.local ceph-mon[47734]: pgmap v1935: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:50 vm08.local ceph-mon[47734]: pgmap v1936: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:50 vm10.local ceph-mon[51861]: pgmap v1936: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:51.571 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:51.571 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:51.597 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:51.597 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:52 vm08.local ceph-mon[47734]: pgmap v1937: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:52 vm10.local ceph-mon[51861]: pgmap v1937: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:54 vm08.local ceph-mon[47734]: pgmap v1938: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:54 vm10.local ceph-mon[51861]: pgmap v1938: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:57:56.599 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:57:56.599 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:57:56.624 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:57:56.624 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:57:57.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:56 vm08.local ceph-mon[47734]: pgmap v1939: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:56 vm10.local ceph-mon[51861]: pgmap v1939: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:58.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:57:58 vm10.local ceph-mon[51861]: pgmap v1940: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:57:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:57:58 vm08.local ceph-mon[47734]: pgmap v1940: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:00 vm08.local ceph-mon[47734]: pgmap v1941: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:00 vm10.local ceph-mon[51861]: pgmap v1941: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:01.626 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:01.626 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:01.670 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:01.671 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:02 vm08.local ceph-mon[47734]: pgmap v1942: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:02 vm10.local ceph-mon[51861]: pgmap v1942: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:04 vm08.local ceph-mon[47734]: pgmap v1943: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:04 vm10.local ceph-mon[51861]: pgmap v1943: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:06.673 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:06.674 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:06.764 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:06.765 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:07.103 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:06 vm08.local ceph-mon[47734]: pgmap v1944: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:06 vm10.local ceph-mon[51861]: pgmap v1944: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:08.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:08 vm08.local ceph-mon[47734]: pgmap v1945: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:08 vm10.local ceph-mon[51861]: pgmap v1945: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:10 vm08.local ceph-mon[47734]: pgmap v1946: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:10 vm10.local ceph-mon[51861]: pgmap v1946: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:11.768 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:11.769 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:11.796 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:11.796 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:13.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:12 vm08.local ceph-mon[47734]: pgmap v1947: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:12 vm10.local ceph-mon[51861]: pgmap v1947: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:14 vm08.local ceph-mon[47734]: pgmap v1948: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:58:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:58:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:58:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:58:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:58:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:14 vm10.local ceph-mon[51861]: pgmap v1948: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:58:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:58:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:58:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:58:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:58:16.798 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:16.798 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:16.825 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:16.826 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:17.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:16 vm08.local ceph-mon[47734]: pgmap v1949: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:16 vm10.local ceph-mon[51861]: pgmap v1949: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:18 vm08.local ceph-mon[47734]: pgmap v1950: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:18 vm10.local ceph-mon[51861]: pgmap v1950: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:20 vm08.local ceph-mon[47734]: pgmap v1951: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:20 vm10.local ceph-mon[51861]: pgmap v1951: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:21.827 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:21.828 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:21.854 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:21.855 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:22 vm08.local ceph-mon[47734]: pgmap v1952: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:22 vm10.local ceph-mon[51861]: pgmap v1952: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:24 vm08.local ceph-mon[47734]: pgmap v1953: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:24 vm10.local ceph-mon[51861]: pgmap v1953: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:26.856 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:26.856 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:26.884 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:26.884 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:27.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:26 vm08.local ceph-mon[47734]: pgmap v1954: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:26 vm10.local ceph-mon[51861]: pgmap v1954: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:28.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:28 vm08.local ceph-mon[47734]: pgmap v1955: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:28 vm10.local ceph-mon[51861]: pgmap v1955: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:31.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:30 vm08.local ceph-mon[47734]: pgmap v1956: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:30 vm10.local ceph-mon[51861]: pgmap v1956: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:31.886 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:31.886 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:31.921 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:31.922 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:32 vm10.local ceph-mon[51861]: pgmap v1957: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:32 vm08.local ceph-mon[47734]: pgmap v1957: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:34 vm10.local ceph-mon[51861]: pgmap v1958: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:34 vm08.local ceph-mon[47734]: pgmap v1958: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:36.923 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:36.924 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:36.952 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:36.953 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:36 vm10.local ceph-mon[51861]: pgmap v1959: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:36 vm08.local ceph-mon[47734]: pgmap v1959: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:38 vm08.local ceph-mon[47734]: pgmap v1960: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:38 vm10.local ceph-mon[51861]: pgmap v1960: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:41.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:40 vm08.local ceph-mon[47734]: pgmap v1961: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:40 vm10.local ceph-mon[51861]: pgmap v1961: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:41.954 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:41.955 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:41.981 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:41.982 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:42 vm08.local ceph-mon[47734]: pgmap v1962: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:42 vm10.local ceph-mon[51861]: pgmap v1962: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:58:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:58:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:58:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:58:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:44 vm08.local ceph-mon[47734]: pgmap v1963: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:44 vm10.local ceph-mon[51861]: pgmap v1963: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:46.983 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:46.984 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:47.044 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:47.045 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:47.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:46 vm08.local ceph-mon[47734]: pgmap v1964: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:46 vm10.local ceph-mon[51861]: pgmap v1964: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:48.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:48 vm08.local ceph-mon[47734]: pgmap v1965: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:48 vm10.local ceph-mon[51861]: pgmap v1965: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:50 vm08.local ceph-mon[47734]: pgmap v1966: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:50 vm10.local ceph-mon[51861]: pgmap v1966: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:52.046 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:52.046 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:52.072 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:52.073 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:53.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:52 vm08.local ceph-mon[47734]: pgmap v1967: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:52 vm10.local ceph-mon[51861]: pgmap v1967: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:54 vm08.local ceph-mon[47734]: pgmap v1968: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:54 vm10.local ceph-mon[51861]: pgmap v1968: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:58:57.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:56 vm08.local ceph-mon[47734]: pgmap v1969: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:57.074 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:58:57.074 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:58:57.099 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:58:57.100 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:58:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:56 vm10.local ceph-mon[51861]: pgmap v1969: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:58:58 vm08.local ceph-mon[47734]: pgmap v1970: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:58:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:58:58 vm10.local ceph-mon[51861]: pgmap v1970: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:00 vm08.local ceph-mon[47734]: pgmap v1971: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:00 vm10.local ceph-mon[51861]: pgmap v1971: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:02.101 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:02.101 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:02.127 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:02.128 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:03.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:02 vm08.local ceph-mon[47734]: pgmap v1972: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:02 vm10.local ceph-mon[51861]: pgmap v1972: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:04 vm08.local ceph-mon[47734]: pgmap v1973: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:04 vm10.local ceph-mon[51861]: pgmap v1973: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:07.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:06 vm08.local ceph-mon[47734]: pgmap v1974: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:07.129 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:07.130 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:06 vm10.local ceph-mon[51861]: pgmap v1974: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:07.170 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:07.171 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:08.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:08 vm08.local ceph-mon[47734]: pgmap v1975: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:08 vm10.local ceph-mon[51861]: pgmap v1975: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:10 vm08.local ceph-mon[47734]: pgmap v1976: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:10 vm10.local ceph-mon[51861]: pgmap v1976: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:12.173 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:12.173 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:12.199 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:12.200 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:12 vm08.local ceph-mon[47734]: pgmap v1977: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:12 vm10.local ceph-mon[51861]: pgmap v1977: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:14 vm08.local ceph-mon[47734]: pgmap v1978: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:59:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:59:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:59:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:59:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:59:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:14 vm10.local ceph-mon[51861]: pgmap v1978: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T14:59:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T14:59:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T14:59:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:59:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T14:59:17.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:16 vm08.local ceph-mon[47734]: pgmap v1979: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:16 vm10.local ceph-mon[51861]: pgmap v1979: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:17.201 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:17.202 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:17.236 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:17.236 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:18 vm08.local ceph-mon[47734]: pgmap v1980: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:18 vm10.local ceph-mon[51861]: pgmap v1980: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:20 vm08.local ceph-mon[47734]: pgmap v1981: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:20 vm10.local ceph-mon[51861]: pgmap v1981: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:22.238 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:22.239 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:22.266 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:22.267 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:22 vm08.local ceph-mon[47734]: pgmap v1982: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:22 vm10.local ceph-mon[51861]: pgmap v1982: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:25.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:24 vm08.local ceph-mon[47734]: pgmap v1983: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:24 vm10.local ceph-mon[51861]: pgmap v1983: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:27.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:26 vm08.local ceph-mon[47734]: pgmap v1984: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:26 vm10.local ceph-mon[51861]: pgmap v1984: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:27.268 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:27.268 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:27.293 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:27.294 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:28.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:28 vm08.local ceph-mon[47734]: pgmap v1985: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:28 vm10.local ceph-mon[51861]: pgmap v1985: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:30 vm08.local ceph-mon[47734]: pgmap v1986: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:30 vm10.local ceph-mon[51861]: pgmap v1986: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:32.295 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:32.295 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:32.321 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:32.322 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:32 vm08.local ceph-mon[47734]: pgmap v1987: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:32 vm10.local ceph-mon[51861]: pgmap v1987: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:34 vm08.local ceph-mon[47734]: pgmap v1988: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:34 vm10.local ceph-mon[51861]: pgmap v1988: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:37.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:36 vm08.local ceph-mon[47734]: pgmap v1989: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:36 vm10.local ceph-mon[51861]: pgmap v1989: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:37.323 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:37.323 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:37.351 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:37.351 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:38 vm08.local ceph-mon[47734]: pgmap v1990: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:38 vm10.local ceph-mon[51861]: pgmap v1990: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:40 vm08.local ceph-mon[47734]: pgmap v1991: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:40 vm10.local ceph-mon[51861]: pgmap v1991: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:42.353 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:42.353 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:42.381 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:42.381 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:42 vm10.local ceph-mon[51861]: pgmap v1992: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:42 vm08.local ceph-mon[47734]: pgmap v1992: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:59:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:59:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T14:59:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T14:59:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:44 vm10.local ceph-mon[51861]: pgmap v1993: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:44 vm08.local ceph-mon[47734]: pgmap v1993: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:46 vm10.local ceph-mon[51861]: pgmap v1994: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:47.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:46 vm08.local ceph-mon[47734]: pgmap v1994: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:47.383 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:47.384 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:47.412 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:47.412 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:48 vm10.local ceph-mon[51861]: pgmap v1995: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:48.960 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:48 vm08.local ceph-mon[47734]: pgmap v1995: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:50 vm08.local ceph-mon[47734]: pgmap v1996: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:50 vm10.local ceph-mon[51861]: pgmap v1996: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:52.413 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:52.414 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:52.441 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:52.442 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:52 vm08.local ceph-mon[47734]: pgmap v1997: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:52 vm10.local ceph-mon[51861]: pgmap v1997: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:55 vm08.local ceph-mon[47734]: pgmap v1998: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:55 vm10.local ceph-mon[51861]: pgmap v1998: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T14:59:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:57 vm08.local ceph-mon[47734]: pgmap v1999: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:57 vm10.local ceph-mon[51861]: pgmap v1999: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:57.443 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T14:59:57.444 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T14:59:57.544 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T14:59:57.545 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T14:59:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 14:59:58 vm10.local ceph-mon[51861]: pgmap v2000: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T14:59:58.962 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 14:59:58 vm08.local ceph-mon[47734]: pgmap v2000: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:00 vm10.local ceph-mon[51861]: pgmap v2001: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T15:00:01.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:00 vm08.local ceph-mon[47734]: pgmap v2001: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:01.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T15:00:02.546 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:02.547 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:02.574 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:02.575 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:02 vm10.local ceph-mon[51861]: pgmap v2002: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:02 vm08.local ceph-mon[47734]: pgmap v2002: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:04 vm10.local ceph-mon[51861]: pgmap v2003: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:04 vm08.local ceph-mon[47734]: pgmap v2003: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:07.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:06 vm08.local ceph-mon[47734]: pgmap v2004: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:06 vm10.local ceph-mon[51861]: pgmap v2004: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:07.576 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:07.576 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:07.602 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:07.602 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:08.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:08 vm08.local ceph-mon[47734]: pgmap v2005: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:08 vm10.local ceph-mon[51861]: pgmap v2005: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:10 vm08.local ceph-mon[47734]: pgmap v2006: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:10 vm10.local ceph-mon[51861]: pgmap v2006: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:12.604 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:12.605 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:12.632 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:12.632 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:12 vm08.local ceph-mon[47734]: pgmap v2007: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:12 vm10.local ceph-mon[51861]: pgmap v2007: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:14 vm08.local ceph-mon[47734]: pgmap v2008: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:00:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:00:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:14 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:00:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:14 vm10.local ceph-mon[51861]: pgmap v2008: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:00:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:00:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:14 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:00:16.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:00:16.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:00:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:00:16.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:00:17.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:16 vm08.local ceph-mon[47734]: pgmap v2009: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:16 vm10.local ceph-mon[51861]: pgmap v2009: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:17.634 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:17.634 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:17.662 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:17.663 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:18.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:18 vm08.local ceph-mon[47734]: pgmap v2010: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:18 vm10.local ceph-mon[51861]: pgmap v2010: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:20 vm08.local ceph-mon[47734]: pgmap v2011: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:20 vm10.local ceph-mon[51861]: pgmap v2011: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:22.665 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:22.665 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:22.696 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:22.696 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:23.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:22 vm08.local ceph-mon[47734]: pgmap v2012: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:22 vm10.local ceph-mon[51861]: pgmap v2012: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:24 vm08.local ceph-mon[47734]: pgmap v2013: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:24 vm10.local ceph-mon[51861]: pgmap v2013: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:26 vm10.local ceph-mon[51861]: pgmap v2014: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:26 vm08.local ceph-mon[47734]: pgmap v2014: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:27.698 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:27.699 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:27.725 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:27.726 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:28.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:28 vm10.local ceph-mon[51861]: pgmap v2015: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:28.965 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:28 vm08.local ceph-mon[47734]: pgmap v2015: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:31.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:30 vm08.local ceph-mon[47734]: pgmap v2016: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:30 vm10.local ceph-mon[51861]: pgmap v2016: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:32.727 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:32.728 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:32.753 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:32.753 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:32 vm10.local ceph-mon[51861]: pgmap v2017: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:32 vm08.local ceph-mon[47734]: pgmap v2017: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:35.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:34 vm08.local ceph-mon[47734]: pgmap v2018: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:34 vm10.local ceph-mon[51861]: pgmap v2018: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:37.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:36 vm08.local ceph-mon[47734]: pgmap v2019: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:36 vm10.local ceph-mon[51861]: pgmap v2019: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:37.755 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:37.755 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:37.784 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:37.785 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:39.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:38 vm08.local ceph-mon[47734]: pgmap v2020: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:38 vm10.local ceph-mon[51861]: pgmap v2020: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:40 vm08.local ceph-mon[47734]: pgmap v2021: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:40 vm10.local ceph-mon[51861]: pgmap v2021: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:42.786 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:42.787 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:42.815 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:42.815 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:42 vm10.local ceph-mon[51861]: pgmap v2022: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:43.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:42 vm08.local ceph-mon[47734]: pgmap v2022: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:00:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:00:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:00:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:00:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:44 vm10.local ceph-mon[51861]: pgmap v2023: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:45.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:44 vm08.local ceph-mon[47734]: pgmap v2023: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:46 vm08.local ceph-mon[47734]: pgmap v2024: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:46 vm10.local ceph-mon[51861]: pgmap v2024: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:47.817 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:47.818 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:47.859 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:47.860 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:48.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:48 vm10.local ceph-mon[51861]: pgmap v2025: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:48.967 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:48 vm08.local ceph-mon[47734]: pgmap v2025: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:50 vm08.local ceph-mon[47734]: pgmap v2026: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:50 vm10.local ceph-mon[51861]: pgmap v2026: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:52.862 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:52.863 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:52.897 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:52.897 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:52 vm08.local ceph-mon[47734]: pgmap v2027: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:52 vm10.local ceph-mon[51861]: pgmap v2027: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:54 vm10.local ceph-mon[51861]: pgmap v2028: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:55.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:54 vm08.local ceph-mon[47734]: pgmap v2028: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:00:57.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:56 vm08.local ceph-mon[47734]: pgmap v2029: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:56 vm10.local ceph-mon[51861]: pgmap v2029: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:57.899 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:00:57.899 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:00:57.988 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:00:57.988 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:00:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:00:58 vm10.local ceph-mon[51861]: pgmap v2030: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:00:58.968 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:00:58 vm08.local ceph-mon[47734]: pgmap v2030: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:01.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:00 vm08.local ceph-mon[47734]: pgmap v2031: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:00 vm10.local ceph-mon[51861]: pgmap v2031: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:02.992 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:02.992 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:03.031 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:03.032 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:02 vm08.local ceph-mon[47734]: pgmap v2032: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:02 vm10.local ceph-mon[51861]: pgmap v2032: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:04 vm08.local ceph-mon[47734]: pgmap v2033: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:04 vm10.local ceph-mon[51861]: pgmap v2033: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:07.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:06 vm08.local ceph-mon[47734]: pgmap v2034: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:06 vm10.local ceph-mon[51861]: pgmap v2034: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:08.034 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:08.035 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:08.062 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:08.063 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:08.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:08 vm10.local ceph-mon[51861]: pgmap v2035: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:08.969 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:08 vm08.local ceph-mon[47734]: pgmap v2035: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:10 vm08.local ceph-mon[47734]: pgmap v2036: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:10 vm10.local ceph-mon[51861]: pgmap v2036: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:13.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:12 vm08.local ceph-mon[47734]: pgmap v2037: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:13.064 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:13.064 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:13.102 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:13.102 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:12 vm10.local ceph-mon[51861]: pgmap v2037: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:15.048 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:14 vm08.local ceph-mon[47734]: pgmap v2038: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:14 vm10.local ceph-mon[51861]: pgmap v2038: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:16.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:01:16.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:01:16.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:01:16.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:01:16.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:01:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:01:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:01:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:01:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:01:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:01:17.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:17 vm08.local ceph-mon[47734]: pgmap v2039: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:17 vm10.local ceph-mon[51861]: pgmap v2039: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:18.104 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:18.104 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:18.130 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:18.131 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:18 vm10.local ceph-mon[51861]: pgmap v2040: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:18.970 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:18 vm08.local ceph-mon[47734]: pgmap v2040: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:20 vm08.local ceph-mon[47734]: pgmap v2041: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:20 vm10.local ceph-mon[51861]: pgmap v2041: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:23.132 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:23.133 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:22 vm08.local ceph-mon[47734]: pgmap v2042: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:23.378 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:23.379 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:22 vm10.local ceph-mon[51861]: pgmap v2042: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:25.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:25 vm08.local ceph-mon[47734]: pgmap v2043: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:25 vm10.local ceph-mon[51861]: pgmap v2043: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:27 vm08.local ceph-mon[47734]: pgmap v2044: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:27 vm10.local ceph-mon[51861]: pgmap v2044: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:28.380 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:28.381 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:28.484 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:28.485 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:28 vm10.local ceph-mon[51861]: pgmap v2045: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:29.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:28 vm08.local ceph-mon[47734]: pgmap v2045: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:30 vm10.local ceph-mon[51861]: pgmap v2046: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:31.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:30 vm08.local ceph-mon[47734]: pgmap v2046: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:33.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:33 vm08.local ceph-mon[47734]: pgmap v2047: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:33 vm10.local ceph-mon[51861]: pgmap v2047: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:33.487 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:33.487 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:33.577 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:33.577 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:35.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:35 vm08.local ceph-mon[47734]: pgmap v2048: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:35 vm10.local ceph-mon[51861]: pgmap v2048: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:37 vm08.local ceph-mon[47734]: pgmap v2049: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:37 vm10.local ceph-mon[51861]: pgmap v2049: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:38.579 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:38.579 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:38.606 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:38.607 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:38.940 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:38 vm10.local ceph-mon[51861]: pgmap v2050: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:38.971 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:38 vm08.local ceph-mon[47734]: pgmap v2050: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:40 vm08.local ceph-mon[47734]: pgmap v2051: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:40 vm10.local ceph-mon[51861]: pgmap v2051: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:42 vm08.local ceph-mon[47734]: pgmap v2052: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:42 vm10.local ceph-mon[51861]: pgmap v2052: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:43.608 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:43.609 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:43.652 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:43.652 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:01:44.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:01:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:01:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:01:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:44 vm10.local ceph-mon[51861]: pgmap v2053: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:45.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:44 vm08.local ceph-mon[47734]: pgmap v2053: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:46 vm10.local ceph-mon[51861]: pgmap v2054: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:47.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:46 vm08.local ceph-mon[47734]: pgmap v2054: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:48.654 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:48.655 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:48.681 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:48.682 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:48.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:48 vm10.local ceph-mon[51861]: pgmap v2055: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:48.972 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:48 vm08.local ceph-mon[47734]: pgmap v2055: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:50 vm08.local ceph-mon[47734]: pgmap v2056: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:50 vm10.local ceph-mon[51861]: pgmap v2056: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:01:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:52 vm10.local ceph-mon[51861]: pgmap v2057: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:53.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:52 vm08.local ceph-mon[47734]: pgmap v2057: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:53.683 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:53.684 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:53.710 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:53.711 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:54 vm10.local ceph-mon[51861]: pgmap v2058: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:55.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:54 vm08.local ceph-mon[47734]: pgmap v2058: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:01:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:56 vm10.local ceph-mon[51861]: pgmap v2059: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:01:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:56 vm08.local ceph-mon[47734]: pgmap v2059: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:01:58.712 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:01:58.713 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:01:58.738 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:01:58.738 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:01:58.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:01:58 vm10.local ceph-mon[51861]: pgmap v2060: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:01:58.973 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:01:58 vm08.local ceph-mon[47734]: pgmap v2060: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:02:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:00 vm08.local ceph-mon[47734]: pgmap v2061: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:00 vm10.local ceph-mon[51861]: pgmap v2061: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:03.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:02 vm08.local ceph-mon[47734]: pgmap v2062: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:02:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:02 vm10.local ceph-mon[51861]: pgmap v2062: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:02:03.740 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:03.741 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:03.782 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:03.782 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:05.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:04 vm08.local ceph-mon[47734]: pgmap v2063: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:04 vm10.local ceph-mon[51861]: pgmap v2063: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:07.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:06 vm10.local ceph-mon[51861]: pgmap v2064: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:06 vm08.local ceph-mon[47734]: pgmap v2064: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:08.784 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:08.784 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:08.815 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:08.815 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:08.940 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:08 vm10.local ceph-mon[51861]: pgmap v2065: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:08.974 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:08 vm08.local ceph-mon[47734]: pgmap v2065: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:10 vm08.local ceph-mon[47734]: pgmap v2066: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:10 vm10.local ceph-mon[51861]: pgmap v2066: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:13.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:12 vm08.local ceph-mon[47734]: pgmap v2067: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:13.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:12 vm10.local ceph-mon[51861]: pgmap v2067: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:13.817 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:13.817 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:13.845 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:13.845 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:15.088 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:14 vm08.local ceph-mon[47734]: pgmap v2068: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:15.090 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:14 vm10.local ceph-mon[51861]: pgmap v2068: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:02:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:02:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:02:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:02:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:02:16.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:02:16.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:02:16.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:02:16.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:02:16.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:02:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:16 vm10.local ceph-mon[51861]: pgmap v2069: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:16 vm08.local ceph-mon[47734]: pgmap v2069: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:18.847 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:18.847 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:18.878 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:18.879 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:18.939 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:18 vm10.local ceph-mon[51861]: pgmap v2070: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:18.975 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:18 vm08.local ceph-mon[47734]: pgmap v2070: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:20 vm08.local ceph-mon[47734]: pgmap v2071: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:20 vm10.local ceph-mon[51861]: pgmap v2071: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:22 vm08.local ceph-mon[47734]: pgmap v2072: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:22 vm10.local ceph-mon[51861]: pgmap v2072: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:23.880 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:23.881 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:23.909 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:23.909 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:25.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:24 vm08.local ceph-mon[47734]: pgmap v2073: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:24 vm10.local ceph-mon[51861]: pgmap v2073: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:26 vm10.local ceph-mon[51861]: pgmap v2074: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:26 vm08.local ceph-mon[47734]: pgmap v2074: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:28.911 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:28.911 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:28.940 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:28 vm10.local ceph-mon[51861]: pgmap v2075: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:28.941 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:28.942 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:28.976 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:28 vm08.local ceph-mon[47734]: pgmap v2075: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:30 vm08.local ceph-mon[47734]: pgmap v2076: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:30 vm10.local ceph-mon[51861]: pgmap v2076: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:32 vm08.local ceph-mon[47734]: pgmap v2077: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:32 vm10.local ceph-mon[51861]: pgmap v2077: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:33.943 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:33.944 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:33.974 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:33.974 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:34 vm08.local ceph-mon[47734]: pgmap v2078: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:34 vm10.local ceph-mon[51861]: pgmap v2078: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:36 vm10.local ceph-mon[51861]: pgmap v2079: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:36 vm08.local ceph-mon[47734]: pgmap v2079: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:38.976 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:38.976 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:38 vm08.local ceph-mon[47734]: pgmap v2080: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:38.976 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:39.004 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:39.004 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:38 vm10.local ceph-mon[51861]: pgmap v2080: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:41.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:40 vm08.local ceph-mon[47734]: pgmap v2081: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:40 vm10.local ceph-mon[51861]: pgmap v2081: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:43.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:42 vm08.local ceph-mon[47734]: pgmap v2082: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:42 vm10.local ceph-mon[51861]: pgmap v2082: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:44.005 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:44.006 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:02:44.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:02:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:02:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:02:44.165 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:44.166 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:45.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:44 vm08.local ceph-mon[47734]: pgmap v2083: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:44 vm10.local ceph-mon[51861]: pgmap v2083: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:46 vm10.local ceph-mon[51861]: pgmap v2084: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:46 vm08.local ceph-mon[47734]: pgmap v2084: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:48.977 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:48 vm08.local ceph-mon[47734]: pgmap v2085: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:48 vm10.local ceph-mon[51861]: pgmap v2085: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:49.167 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:49.167 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:49.261 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:49.262 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:51.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:50 vm08.local ceph-mon[47734]: pgmap v2086: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:50 vm10.local ceph-mon[51861]: pgmap v2086: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:53.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:52 vm08.local ceph-mon[47734]: pgmap v2087: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:52 vm10.local ceph-mon[51861]: pgmap v2087: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:54.263 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:54.263 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:54.289 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:54.290 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:02:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:54 vm10.local ceph-mon[51861]: pgmap v2088: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:54 vm08.local ceph-mon[47734]: pgmap v2088: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:02:57.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:56 vm10.local ceph-mon[51861]: pgmap v2089: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:57.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:56 vm08.local ceph-mon[47734]: pgmap v2089: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:58.978 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:02:58 vm08.local ceph-mon[47734]: pgmap v2090: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:02:58 vm10.local ceph-mon[51861]: pgmap v2090: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:02:59.291 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:02:59.292 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:02:59.317 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:02:59.317 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:00 vm10.local ceph-mon[51861]: pgmap v2091: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:00 vm08.local ceph-mon[47734]: pgmap v2091: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:03.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:02 vm10.local ceph-mon[51861]: pgmap v2092: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:02 vm08.local ceph-mon[47734]: pgmap v2092: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:04.319 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:04.319 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:04.344 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:04.345 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:05.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:04 vm08.local ceph-mon[47734]: pgmap v2093: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:04 vm10.local ceph-mon[51861]: pgmap v2093: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:07.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:06 vm08.local ceph-mon[47734]: pgmap v2094: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:06 vm10.local ceph-mon[51861]: pgmap v2094: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:08.980 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:08 vm08.local ceph-mon[47734]: pgmap v2095: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:09.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:08 vm10.local ceph-mon[51861]: pgmap v2095: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:09.346 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:09.347 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:09.471 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:09.472 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:11.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:10 vm08.local ceph-mon[47734]: pgmap v2096: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:10 vm10.local ceph-mon[51861]: pgmap v2096: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:13.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:12 vm08.local ceph-mon[47734]: pgmap v2097: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:12 vm10.local ceph-mon[51861]: pgmap v2097: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:14.474 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:14.474 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:14.503 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:14.504 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:14 vm08.local ceph-mon[47734]: pgmap v2098: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:15.364 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:14 vm10.local ceph-mon[51861]: pgmap v2098: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:16.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:03:16.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:03:16.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:03:16.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:03:16.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:03:16.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:03:16.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:03:16.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:03:16.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:03:16.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:03:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:16 vm08.local ceph-mon[47734]: pgmap v2099: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:16 vm10.local ceph-mon[51861]: pgmap v2099: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:19 vm08.local ceph-mon[47734]: pgmap v2100: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:19 vm10.local ceph-mon[51861]: pgmap v2100: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:19.505 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:19.505 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:19.531 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:19.532 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:21 vm10.local ceph-mon[51861]: pgmap v2101: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:21 vm08.local ceph-mon[47734]: pgmap v2101: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:23.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:23 vm08.local ceph-mon[47734]: pgmap v2102: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:23 vm10.local ceph-mon[51861]: pgmap v2102: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:24.534 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:24.534 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:24 vm08.local ceph-mon[47734]: pgmap v2103: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:24.561 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:24.562 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:24 vm10.local ceph-mon[51861]: pgmap v2103: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:27.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:26 vm08.local ceph-mon[47734]: pgmap v2104: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:26 vm10.local ceph-mon[51861]: pgmap v2104: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:29.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:28 vm08.local ceph-mon[47734]: pgmap v2105: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:28 vm10.local ceph-mon[51861]: pgmap v2105: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:29.564 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:29.564 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:29.591 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:29.592 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:30 vm10.local ceph-mon[51861]: pgmap v2106: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:30 vm08.local ceph-mon[47734]: pgmap v2106: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:32 vm10.local ceph-mon[51861]: pgmap v2107: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:32 vm08.local ceph-mon[47734]: pgmap v2107: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:34.593 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:34.593 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:34.682 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:34.683 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:34 vm10.local ceph-mon[51861]: pgmap v2108: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:35.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:34 vm08.local ceph-mon[47734]: pgmap v2108: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:37.062 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:36 vm08.local ceph-mon[47734]: pgmap v2109: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:36 vm10.local ceph-mon[51861]: pgmap v2109: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:38 vm10.local ceph-mon[51861]: pgmap v2110: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:38 vm08.local ceph-mon[47734]: pgmap v2110: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:39.685 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:39.685 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:39.713 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:39.713 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:41.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:40 vm10.local ceph-mon[51861]: pgmap v2111: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:40 vm08.local ceph-mon[47734]: pgmap v2111: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:42 vm10.local ceph-mon[51861]: pgmap v2112: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:42 vm08.local ceph-mon[47734]: pgmap v2112: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:03:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:03:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:03:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:03:44.714 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:44.715 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:44.741 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:44.741 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:44 vm10.local ceph-mon[51861]: pgmap v2113: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:44 vm08.local ceph-mon[47734]: pgmap v2113: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:47.208 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:46 vm08.local ceph-mon[47734]: pgmap v2114: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:46 vm10.local ceph-mon[51861]: pgmap v2114: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:48 vm08.local ceph-mon[47734]: pgmap v2115: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:48 vm10.local ceph-mon[51861]: pgmap v2115: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:49.743 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:49.743 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:49.770 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:49.770 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:51.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:51 vm08.local ceph-mon[47734]: pgmap v2116: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:51 vm10.local ceph-mon[51861]: pgmap v2116: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:53 vm08.local ceph-mon[47734]: pgmap v2117: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:53 vm10.local ceph-mon[51861]: pgmap v2117: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:54.772 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:54.773 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:54.801 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:54.802 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:03:55.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:55 vm08.local ceph-mon[47734]: pgmap v2118: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:55 vm10.local ceph-mon[51861]: pgmap v2118: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:03:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:57 vm08.local ceph-mon[47734]: pgmap v2119: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:57 vm10.local ceph-mon[51861]: pgmap v2119: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:03:59 vm08.local ceph-mon[47734]: pgmap v2120: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:03:59 vm10.local ceph-mon[51861]: pgmap v2120: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:03:59.804 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:03:59.804 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:03:59.830 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:03:59.831 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:01 vm10.local ceph-mon[51861]: pgmap v2121: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:01 vm08.local ceph-mon[47734]: pgmap v2121: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:03 vm10.local ceph-mon[51861]: pgmap v2122: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:03 vm08.local ceph-mon[47734]: pgmap v2122: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:04.832 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:04.833 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:04.859 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:04.859 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:05 vm10.local ceph-mon[51861]: pgmap v2123: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:05.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:05 vm08.local ceph-mon[47734]: pgmap v2123: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:07 vm10.local ceph-mon[51861]: pgmap v2124: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:07 vm08.local ceph-mon[47734]: pgmap v2124: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:09.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:08 vm08.local ceph-mon[47734]: pgmap v2125: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:09.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:08 vm10.local ceph-mon[51861]: pgmap v2125: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:09.861 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:09.861 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:09.891 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:09.891 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:10 vm08.local ceph-mon[47734]: pgmap v2126: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:10 vm10.local ceph-mon[51861]: pgmap v2126: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:12 vm10.local ceph-mon[51861]: pgmap v2127: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:12 vm08.local ceph-mon[47734]: pgmap v2127: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:14.893 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:14.893 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:14.919 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:14.920 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:14 vm10.local ceph-mon[51861]: pgmap v2128: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:14 vm08.local ceph-mon[47734]: pgmap v2128: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:15.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:04:15.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:04:15.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:15 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:04:16.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:04:16.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:04:16.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:15 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:04:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:16 vm10.local ceph-mon[51861]: pgmap v2129: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:16 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:04:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:16 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:04:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:16 vm08.local ceph-mon[47734]: pgmap v2129: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:16 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:04:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:16 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:04:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:18 vm10.local ceph-mon[51861]: pgmap v2130: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:18 vm08.local ceph-mon[47734]: pgmap v2130: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:19.922 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:19.922 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:19.957 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:19.957 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:21.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:20 vm08.local ceph-mon[47734]: pgmap v2131: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:20 vm10.local ceph-mon[51861]: pgmap v2131: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:23.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:22 vm08.local ceph-mon[47734]: pgmap v2132: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:22 vm10.local ceph-mon[51861]: pgmap v2132: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:24.959 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:24.959 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:24.984 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:24.985 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:24 vm08.local ceph-mon[47734]: pgmap v2133: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:24 vm10.local ceph-mon[51861]: pgmap v2133: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:27 vm10.local ceph-mon[51861]: pgmap v2134: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:27 vm08.local ceph-mon[47734]: pgmap v2134: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:29 vm08.local ceph-mon[47734]: pgmap v2135: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:29 vm10.local ceph-mon[51861]: pgmap v2135: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:29.986 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:29.987 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:30.015 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:30.015 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:31 vm10.local ceph-mon[51861]: pgmap v2136: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:31 vm08.local ceph-mon[47734]: pgmap v2136: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:33 vm10.local ceph-mon[51861]: pgmap v2137: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:33.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:33 vm08.local ceph-mon[47734]: pgmap v2137: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:34 vm08.local ceph-mon[47734]: pgmap v2138: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:34 vm10.local ceph-mon[51861]: pgmap v2138: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:35.017 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:35.017 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:35.043 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:35.043 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:36 vm10.local ceph-mon[51861]: pgmap v2139: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:37.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:36 vm08.local ceph-mon[47734]: pgmap v2139: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:38 vm10.local ceph-mon[51861]: pgmap v2140: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:38 vm08.local ceph-mon[47734]: pgmap v2140: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:40.044 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:40.045 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:40.071 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:40.071 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:40 vm10.local ceph-mon[51861]: pgmap v2141: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:41.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:40 vm08.local ceph-mon[47734]: pgmap v2141: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:43.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:42 vm10.local ceph-mon[51861]: pgmap v2142: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:42 vm08.local ceph-mon[47734]: pgmap v2142: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:04:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:04:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:04:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:04:45.073 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:45.073 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:45.098 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:45.099 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:44 vm08.local ceph-mon[47734]: pgmap v2143: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:44 vm10.local ceph-mon[51861]: pgmap v2143: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:46 vm08.local ceph-mon[47734]: pgmap v2144: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:46 vm10.local ceph-mon[51861]: pgmap v2144: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:48 vm08.local ceph-mon[47734]: pgmap v2145: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:48 vm10.local ceph-mon[51861]: pgmap v2145: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:50.100 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:50.100 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:50.127 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:50.128 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:51.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:51 vm08.local ceph-mon[47734]: pgmap v2146: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:51 vm10.local ceph-mon[51861]: pgmap v2146: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:53 vm08.local ceph-mon[47734]: pgmap v2147: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:53 vm10.local ceph-mon[51861]: pgmap v2147: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:55.130 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:04:55.130 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:04:55.156 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:04:55.156 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:04:55.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:55 vm08.local ceph-mon[47734]: pgmap v2148: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:55 vm10.local ceph-mon[51861]: pgmap v2148: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:04:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:57 vm08.local ceph-mon[47734]: pgmap v2149: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:57 vm10.local ceph-mon[51861]: pgmap v2149: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:04:59 vm08.local ceph-mon[47734]: pgmap v2150: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:04:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:04:59 vm10.local ceph-mon[51861]: pgmap v2150: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:00.158 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:00.158 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:00.185 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:00.185 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:01 vm10.local ceph-mon[51861]: pgmap v2151: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:01.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:01 vm08.local ceph-mon[47734]: pgmap v2151: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:03 vm10.local ceph-mon[51861]: pgmap v2152: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:03.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:03 vm08.local ceph-mon[47734]: pgmap v2152: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:05.187 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:05.187 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:05.213 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:05.213 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:05 vm10.local ceph-mon[51861]: pgmap v2153: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:05 vm08.local ceph-mon[47734]: pgmap v2153: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:07 vm10.local ceph-mon[51861]: pgmap v2154: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:07 vm08.local ceph-mon[47734]: pgmap v2154: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:09 vm08.local ceph-mon[47734]: pgmap v2155: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:09 vm10.local ceph-mon[51861]: pgmap v2155: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:10.215 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:10.215 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:10.241 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:10.242 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:11 vm10.local ceph-mon[51861]: pgmap v2156: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:11 vm08.local ceph-mon[47734]: pgmap v2156: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:13 vm10.local ceph-mon[51861]: pgmap v2157: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:13 vm08.local ceph-mon[47734]: pgmap v2157: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:15.243 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:15.243 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:15.269 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:15.269 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:15 vm10.local ceph-mon[51861]: pgmap v2158: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:15 vm08.local ceph-mon[47734]: pgmap v2158: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:16.139 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:16 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:05:16.139 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:16 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:05:16.139 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:16 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:05:16.139 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:16 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:05:16.139 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:16 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:05:16.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:16 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:05:16.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:16 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:05:16.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:16 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:05:16.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:16 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:05:16.159 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:16 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:05:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:17 vm10.local ceph-mon[51861]: pgmap v2159: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:05:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:05:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:17 vm08.local ceph-mon[47734]: pgmap v2159: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:05:17.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:05:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:19 vm08.local ceph-mon[47734]: pgmap v2160: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:19 vm10.local ceph-mon[51861]: pgmap v2160: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:20.271 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:20.271 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:20.297 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:20.297 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:21 vm10.local ceph-mon[51861]: pgmap v2161: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:21.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:21 vm08.local ceph-mon[47734]: pgmap v2161: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:23 vm10.local ceph-mon[51861]: pgmap v2162: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:23 vm08.local ceph-mon[47734]: pgmap v2162: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:25.299 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:25.299 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:25.327 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:25.327 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:25 vm10.local ceph-mon[51861]: pgmap v2163: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:25 vm08.local ceph-mon[47734]: pgmap v2163: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:27 vm10.local ceph-mon[51861]: pgmap v2164: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:27 vm08.local ceph-mon[47734]: pgmap v2164: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:29.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:28 vm10.local ceph-mon[51861]: pgmap v2165: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:28 vm08.local ceph-mon[47734]: pgmap v2165: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:30.329 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:30.329 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:30.355 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:30.356 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:30 vm10.local ceph-mon[51861]: pgmap v2166: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:31.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:30 vm08.local ceph-mon[47734]: pgmap v2166: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:32 vm10.local ceph-mon[51861]: pgmap v2167: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:32 vm08.local ceph-mon[47734]: pgmap v2167: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:34 vm10.local ceph-mon[51861]: pgmap v2168: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:34 vm08.local ceph-mon[47734]: pgmap v2168: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:35.357 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:35.357 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:35.383 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:35.384 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:36 vm10.local ceph-mon[51861]: pgmap v2169: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:36 vm08.local ceph-mon[47734]: pgmap v2169: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:39.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:38 vm08.local ceph-mon[47734]: pgmap v2170: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:38 vm10.local ceph-mon[51861]: pgmap v2170: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:40.385 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:40.385 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:40.413 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:40.413 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:41.210 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:40 vm10.local ceph-mon[51861]: pgmap v2171: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:40 vm08.local ceph-mon[47734]: pgmap v2171: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:42 vm08.local ceph-mon[47734]: pgmap v2172: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:42 vm10.local ceph-mon[51861]: pgmap v2172: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:05:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:05:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:05:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:05:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:44 vm08.local ceph-mon[47734]: pgmap v2173: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:44 vm10.local ceph-mon[51861]: pgmap v2173: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:45.414 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:45.415 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:45.445 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:45.446 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:47.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:46 vm08.local ceph-mon[47734]: pgmap v2174: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:46 vm10.local ceph-mon[51861]: pgmap v2174: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:48 vm08.local ceph-mon[47734]: pgmap v2175: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:48 vm10.local ceph-mon[51861]: pgmap v2175: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:50.447 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:50.448 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:50.476 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:50.477 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:50 vm08.local ceph-mon[47734]: pgmap v2176: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:50 vm10.local ceph-mon[51861]: pgmap v2176: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:53.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:52 vm08.local ceph-mon[47734]: pgmap v2177: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:52 vm10.local ceph-mon[51861]: pgmap v2177: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:55.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:54 vm08.local ceph-mon[47734]: pgmap v2178: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:54 vm10.local ceph-mon[51861]: pgmap v2178: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:05:55.478 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:05:55.478 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:05:55.505 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:05:55.506 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:05:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:56 vm08.local ceph-mon[47734]: pgmap v2179: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:56 vm10.local ceph-mon[51861]: pgmap v2179: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:59.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:05:58 vm08.local ceph-mon[47734]: pgmap v2180: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:05:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:05:58 vm10.local ceph-mon[51861]: pgmap v2180: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:00.507 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:00.508 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:00.533 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:00.533 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:00 vm08.local ceph-mon[47734]: pgmap v2181: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:00 vm10.local ceph-mon[51861]: pgmap v2181: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:02 vm08.local ceph-mon[47734]: pgmap v2182: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:02 vm10.local ceph-mon[51861]: pgmap v2182: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:04 vm08.local ceph-mon[47734]: pgmap v2183: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:04 vm10.local ceph-mon[51861]: pgmap v2183: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:05.535 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:05.535 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:05.562 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:05.562 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:06 vm08.local ceph-mon[47734]: pgmap v2184: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:06 vm10.local ceph-mon[51861]: pgmap v2184: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:08 vm08.local ceph-mon[47734]: pgmap v2185: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:08 vm10.local ceph-mon[51861]: pgmap v2185: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:10.564 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:10.564 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:10.590 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:10.591 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:10 vm08.local ceph-mon[47734]: pgmap v2186: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:10 vm10.local ceph-mon[51861]: pgmap v2186: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:13 vm08.local ceph-mon[47734]: pgmap v2187: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:13 vm10.local ceph-mon[51861]: pgmap v2187: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:15 vm08.local ceph-mon[47734]: pgmap v2188: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:15 vm10.local ceph-mon[51861]: pgmap v2188: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:15.592 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:15.592 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:15.618 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:15.619 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:17 vm08.local ceph-mon[47734]: pgmap v2189: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:06:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:06:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:06:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:06:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:06:17.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:06:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:17 vm10.local ceph-mon[51861]: pgmap v2189: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:06:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:06:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:06:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:06:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:06:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:06:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:19 vm08.local ceph-mon[47734]: pgmap v2190: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:19 vm10.local ceph-mon[51861]: pgmap v2190: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:20.620 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:20.620 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:20.648 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:20.648 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:21 vm08.local ceph-mon[47734]: pgmap v2191: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:21 vm10.local ceph-mon[51861]: pgmap v2191: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:23 vm08.local ceph-mon[47734]: pgmap v2192: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:23 vm10.local ceph-mon[51861]: pgmap v2192: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:25 vm08.local ceph-mon[47734]: pgmap v2193: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:25 vm10.local ceph-mon[51861]: pgmap v2193: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:25.649 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:25.650 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:25.677 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:25.678 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:27 vm08.local ceph-mon[47734]: pgmap v2194: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:27 vm10.local ceph-mon[51861]: pgmap v2194: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:29.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:29 vm08.local ceph-mon[47734]: pgmap v2195: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:29 vm10.local ceph-mon[51861]: pgmap v2195: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:30.679 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:30.680 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:30.706 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:30.706 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:31 vm10.local ceph-mon[51861]: pgmap v2196: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:31 vm08.local ceph-mon[47734]: pgmap v2196: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:33 vm10.local ceph-mon[51861]: pgmap v2197: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:33.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:33 vm08.local ceph-mon[47734]: pgmap v2197: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:35 vm10.local ceph-mon[51861]: pgmap v2198: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:35.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:35 vm08.local ceph-mon[47734]: pgmap v2198: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:35.708 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:35.708 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:35.736 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:35.736 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:36.458 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:36 vm08.local ceph-mon[47734]: pgmap v2199: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:36 vm10.local ceph-mon[51861]: pgmap v2199: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:39.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:38 vm10.local ceph-mon[51861]: pgmap v2200: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:39.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:38 vm08.local ceph-mon[47734]: pgmap v2200: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:40.737 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:40.738 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:40.765 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:40.765 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:40 vm08.local ceph-mon[47734]: pgmap v2201: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:40 vm10.local ceph-mon[51861]: pgmap v2201: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:43.211 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:42 vm10.local ceph-mon[51861]: pgmap v2202: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:43.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:42 vm08.local ceph-mon[47734]: pgmap v2202: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:06:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:06:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:06:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:06:45.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:44 vm08.local ceph-mon[47734]: pgmap v2203: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:44 vm10.local ceph-mon[51861]: pgmap v2203: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:45.767 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:45.767 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:45.845 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:45.845 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:47.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:46 vm08.local ceph-mon[47734]: pgmap v2204: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:46 vm10.local ceph-mon[51861]: pgmap v2204: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:48 vm08.local ceph-mon[47734]: pgmap v2205: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:48 vm10.local ceph-mon[51861]: pgmap v2205: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:50.847 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:50.847 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:50.874 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:50.875 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:50 vm08.local ceph-mon[47734]: pgmap v2206: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:50 vm10.local ceph-mon[51861]: pgmap v2206: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:52 vm08.local ceph-mon[47734]: pgmap v2207: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:52 vm10.local ceph-mon[51861]: pgmap v2207: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:54 vm08.local ceph-mon[47734]: pgmap v2208: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:54 vm10.local ceph-mon[51861]: pgmap v2208: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:06:55.876 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:06:55.877 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:06:55.905 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:06:55.905 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:06:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:56 vm08.local ceph-mon[47734]: pgmap v2209: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:56 vm10.local ceph-mon[51861]: pgmap v2209: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:06:58 vm08.local ceph-mon[47734]: pgmap v2210: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:06:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:06:58 vm10.local ceph-mon[51861]: pgmap v2210: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:00.907 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:00.907 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:00.934 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:00.935 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:01.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:00 vm08.local ceph-mon[47734]: pgmap v2211: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:00 vm10.local ceph-mon[51861]: pgmap v2211: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:03 vm10.local ceph-mon[51861]: pgmap v2212: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:03 vm08.local ceph-mon[47734]: pgmap v2212: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:05 vm10.local ceph-mon[51861]: pgmap v2213: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:05 vm08.local ceph-mon[47734]: pgmap v2213: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:05.936 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:05.936 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:05.962 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:05.963 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:07 vm10.local ceph-mon[51861]: pgmap v2214: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:07.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:07 vm08.local ceph-mon[47734]: pgmap v2214: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:09.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:09 vm08.local ceph-mon[47734]: pgmap v2215: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:09 vm10.local ceph-mon[51861]: pgmap v2215: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:10.964 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:10.965 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:10.991 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:10.992 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:11 vm10.local ceph-mon[51861]: pgmap v2216: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:11 vm08.local ceph-mon[47734]: pgmap v2216: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:13 vm10.local ceph-mon[51861]: pgmap v2217: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:13 vm08.local ceph-mon[47734]: pgmap v2217: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:15 vm10.local ceph-mon[51861]: pgmap v2218: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:15 vm08.local ceph-mon[47734]: pgmap v2218: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:15.993 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:15.993 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:16.031 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:16.032 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:17.104 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:17 vm08.local ceph-mon[47734]: pgmap v2219: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:17.104 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:07:17.104 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:07:17.104 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:17 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:07:17.105 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:17 vm10.local ceph-mon[51861]: pgmap v2219: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:17.105 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:07:17.105 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:07:17.105 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:17 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:07:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:07:18.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:07:18.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:18 vm08.local ceph-mon[47734]: pgmap v2220: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:18.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:07:18.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:07:18.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:07:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:07:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:07:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:18 vm10.local ceph-mon[51861]: pgmap v2220: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:07:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:07:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:07:21.033 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:21.033 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:21.061 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:21.061 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:21.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:20 vm08.local ceph-mon[47734]: pgmap v2221: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:20 vm10.local ceph-mon[51861]: pgmap v2221: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:22 vm08.local ceph-mon[47734]: pgmap v2222: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:22 vm10.local ceph-mon[51861]: pgmap v2222: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:24 vm08.local ceph-mon[47734]: pgmap v2223: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:24 vm10.local ceph-mon[51861]: pgmap v2223: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:26.063 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:26.063 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:26.093 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:26.093 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:26 vm08.local ceph-mon[47734]: pgmap v2224: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:26 vm10.local ceph-mon[51861]: pgmap v2224: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:29 vm08.local ceph-mon[47734]: pgmap v2225: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:29 vm10.local ceph-mon[51861]: pgmap v2225: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:31.095 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:31.095 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:31.124 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:31.124 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:31 vm08.local ceph-mon[47734]: pgmap v2226: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:31 vm10.local ceph-mon[51861]: pgmap v2226: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:33 vm08.local ceph-mon[47734]: pgmap v2227: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:33 vm10.local ceph-mon[51861]: pgmap v2227: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:35 vm08.local ceph-mon[47734]: pgmap v2228: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:35 vm10.local ceph-mon[51861]: pgmap v2228: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:36.126 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:36.126 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:36.152 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:36.153 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:37.058 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:37 vm08.local ceph-mon[47734]: pgmap v2229: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:37 vm10.local ceph-mon[51861]: pgmap v2229: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:39 vm08.local ceph-mon[47734]: pgmap v2230: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:39 vm10.local ceph-mon[51861]: pgmap v2230: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:41.155 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:41.155 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:41.181 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:41.182 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:41 vm08.local ceph-mon[47734]: pgmap v2231: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:41 vm10.local ceph-mon[51861]: pgmap v2231: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:43.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:43 vm08.local ceph-mon[47734]: pgmap v2232: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:43.341 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:43 vm10.local ceph-mon[51861]: pgmap v2232: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:07:44.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:07:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:07:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:07:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:45 vm08.local ceph-mon[47734]: pgmap v2233: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:45 vm10.local ceph-mon[51861]: pgmap v2233: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:46.183 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:46.184 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:46.210 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:46.210 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:47 vm10.local ceph-mon[51861]: pgmap v2234: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:47.493 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:47 vm08.local ceph-mon[47734]: pgmap v2234: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:49 vm08.local ceph-mon[47734]: pgmap v2235: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:49 vm10.local ceph-mon[51861]: pgmap v2235: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:51.211 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:51.212 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:51.237 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:51.237 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:51 vm10.local ceph-mon[51861]: pgmap v2236: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:51 vm08.local ceph-mon[47734]: pgmap v2236: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:53 vm10.local ceph-mon[51861]: pgmap v2237: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:53 vm08.local ceph-mon[47734]: pgmap v2237: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:55 vm10.local ceph-mon[51861]: pgmap v2238: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:55 vm08.local ceph-mon[47734]: pgmap v2238: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:07:56.238 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:07:56.239 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:07:56.265 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:07:56.266 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:07:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:57 vm10.local ceph-mon[51861]: pgmap v2239: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:57 vm08.local ceph-mon[47734]: pgmap v2239: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:07:59 vm08.local ceph-mon[47734]: pgmap v2240: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:07:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:07:59 vm10.local ceph-mon[51861]: pgmap v2240: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:01.267 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:01.267 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:01.292 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:01.293 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:01 vm10.local ceph-mon[51861]: pgmap v2241: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:01 vm08.local ceph-mon[47734]: pgmap v2241: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:03 vm10.local ceph-mon[51861]: pgmap v2242: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:03 vm08.local ceph-mon[47734]: pgmap v2242: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:05 vm10.local ceph-mon[51861]: pgmap v2243: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:05 vm08.local ceph-mon[47734]: pgmap v2243: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:06.294 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:06.295 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:06.324 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:06.325 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:07 vm10.local ceph-mon[51861]: pgmap v2244: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:07.427 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:07 vm08.local ceph-mon[47734]: pgmap v2244: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:09 vm08.local ceph-mon[47734]: pgmap v2245: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:09 vm10.local ceph-mon[51861]: pgmap v2245: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:11.326 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:11.326 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:11.352 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:11.353 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:11 vm10.local ceph-mon[51861]: pgmap v2246: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:11 vm08.local ceph-mon[47734]: pgmap v2246: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:13 vm10.local ceph-mon[51861]: pgmap v2247: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:13 vm08.local ceph-mon[47734]: pgmap v2247: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:15 vm10.local ceph-mon[51861]: pgmap v2248: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:15 vm08.local ceph-mon[47734]: pgmap v2248: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:16.354 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:16.355 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:16.383 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:16.384 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:17 vm10.local ceph-mon[51861]: pgmap v2249: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:17.469 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:17 vm08.local ceph-mon[47734]: pgmap v2249: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:18 vm08.local ceph-mon[47734]: pgmap v2250: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:08:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:08:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:08:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:08:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:18 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:08:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:18 vm10.local ceph-mon[51861]: pgmap v2250: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:08:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:08:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:08:19.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:08:19.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:18 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:08:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:20 vm08.local ceph-mon[47734]: pgmap v2251: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:21.385 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:21.386 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:20 vm10.local ceph-mon[51861]: pgmap v2251: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:21.413 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:21.413 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:23 vm08.local ceph-mon[47734]: pgmap v2252: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:23 vm10.local ceph-mon[51861]: pgmap v2252: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:25 vm08.local ceph-mon[47734]: pgmap v2253: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:25 vm10.local ceph-mon[51861]: pgmap v2253: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:26.414 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:26.415 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:26.441 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:26.441 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:27 vm08.local ceph-mon[47734]: pgmap v2254: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:27 vm10.local ceph-mon[51861]: pgmap v2254: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:29.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:29 vm08.local ceph-mon[47734]: pgmap v2255: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:29 vm10.local ceph-mon[51861]: pgmap v2255: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:31 vm08.local ceph-mon[47734]: pgmap v2256: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:31 vm10.local ceph-mon[51861]: pgmap v2256: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:31.442 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:31.443 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:31.470 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:31.470 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:33 vm08.local ceph-mon[47734]: pgmap v2257: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:33 vm10.local ceph-mon[51861]: pgmap v2257: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:35.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:35 vm08.local ceph-mon[47734]: pgmap v2258: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:35 vm10.local ceph-mon[51861]: pgmap v2258: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:36.472 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:36.472 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:36.499 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:36.499 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:37 vm08.local ceph-mon[47734]: pgmap v2259: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:37 vm10.local ceph-mon[51861]: pgmap v2259: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:39.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:39 vm08.local ceph-mon[47734]: pgmap v2260: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:39 vm10.local ceph-mon[51861]: pgmap v2260: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:41 vm10.local ceph-mon[51861]: pgmap v2261: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:41.501 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:41.501 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:41.528 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:41.529 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:41 vm08.local ceph-mon[47734]: pgmap v2261: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:43.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:43 vm10.local ceph-mon[51861]: pgmap v2262: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:43 vm08.local ceph-mon[47734]: pgmap v2262: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:08:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:08:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:08:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:08:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:45 vm10.local ceph-mon[51861]: pgmap v2263: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:45 vm08.local ceph-mon[47734]: pgmap v2263: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:46.530 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:46.531 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:46.557 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:46.557 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:47 vm10.local ceph-mon[51861]: pgmap v2264: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:47 vm08.local ceph-mon[47734]: pgmap v2264: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:49 vm08.local ceph-mon[47734]: pgmap v2265: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:49 vm10.local ceph-mon[51861]: pgmap v2265: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:51 vm10.local ceph-mon[51861]: pgmap v2266: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:51 vm08.local ceph-mon[47734]: pgmap v2266: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:51.558 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:51.559 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:51.586 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:51.586 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:53 vm10.local ceph-mon[51861]: pgmap v2267: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:53 vm08.local ceph-mon[47734]: pgmap v2267: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:55 vm10.local ceph-mon[51861]: pgmap v2268: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:55 vm08.local ceph-mon[47734]: pgmap v2268: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:08:56.587 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:08:56.588 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:08:56.614 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:08:56.615 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:08:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:57 vm10.local ceph-mon[51861]: pgmap v2269: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:57 vm08.local ceph-mon[47734]: pgmap v2269: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:08:59 vm08.local ceph-mon[47734]: pgmap v2270: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:08:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:08:59 vm10.local ceph-mon[51861]: pgmap v2270: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:01 vm10.local ceph-mon[51861]: pgmap v2271: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:01 vm08.local ceph-mon[47734]: pgmap v2271: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:01.616 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:01.617 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:01.644 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:01.644 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:03 vm10.local ceph-mon[51861]: pgmap v2272: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:03 vm08.local ceph-mon[47734]: pgmap v2272: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:05.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:05 vm08.local ceph-mon[47734]: pgmap v2273: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:05 vm10.local ceph-mon[51861]: pgmap v2273: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:06.646 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:06.646 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:06.673 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:06.674 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:07 vm08.local ceph-mon[47734]: pgmap v2274: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:07 vm10.local ceph-mon[51861]: pgmap v2274: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:09 vm08.local ceph-mon[47734]: pgmap v2275: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:09 vm10.local ceph-mon[51861]: pgmap v2275: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:10 vm08.local ceph-mon[47734]: pgmap v2276: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:10 vm10.local ceph-mon[51861]: pgmap v2276: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:11.675 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:11.676 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:11.702 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:11.703 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:13.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:12 vm10.local ceph-mon[51861]: pgmap v2277: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:12 vm08.local ceph-mon[47734]: pgmap v2277: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:15 vm08.local ceph-mon[47734]: pgmap v2278: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:15 vm10.local ceph-mon[51861]: pgmap v2278: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:16.704 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:16.705 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:16.730 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:16.731 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:17 vm08.local ceph-mon[47734]: pgmap v2279: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:17 vm10.local ceph-mon[51861]: pgmap v2279: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:19 vm08.local ceph-mon[47734]: pgmap v2280: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:09:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:09:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:09:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:09:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:09:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:19 vm10.local ceph-mon[51861]: pgmap v2280: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:09:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:09:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:09:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:09:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:09:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:21 vm08.local ceph-mon[47734]: pgmap v2281: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:21 vm10.local ceph-mon[51861]: pgmap v2281: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:21.732 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:21.733 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:21.778 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:21.779 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:23 vm08.local ceph-mon[47734]: pgmap v2282: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:23 vm10.local ceph-mon[51861]: pgmap v2282: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:25 vm08.local ceph-mon[47734]: pgmap v2283: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:25 vm10.local ceph-mon[51861]: pgmap v2283: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:26.781 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:26.781 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:26.810 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:26.811 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:27 vm08.local ceph-mon[47734]: pgmap v2284: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:27 vm10.local ceph-mon[51861]: pgmap v2284: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:29 vm08.local ceph-mon[47734]: pgmap v2285: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:29 vm10.local ceph-mon[51861]: pgmap v2285: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:31 vm10.local ceph-mon[51861]: pgmap v2286: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:31 vm08.local ceph-mon[47734]: pgmap v2286: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:31.812 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:31.813 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:31.842 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:31.842 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:33 vm10.local ceph-mon[51861]: pgmap v2287: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:33 vm08.local ceph-mon[47734]: pgmap v2287: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:35 vm10.local ceph-mon[51861]: pgmap v2288: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:35 vm08.local ceph-mon[47734]: pgmap v2288: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:36.843 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:36.844 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:36.871 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:36.872 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:37 vm10.local ceph-mon[51861]: pgmap v2289: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:37 vm08.local ceph-mon[47734]: pgmap v2289: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:39.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:39 vm08.local ceph-mon[47734]: pgmap v2290: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:39 vm10.local ceph-mon[51861]: pgmap v2290: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:41 vm10.local ceph-mon[51861]: pgmap v2291: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:41.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:41 vm08.local ceph-mon[47734]: pgmap v2291: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:41.873 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:41.874 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:41.905 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:41.905 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:43 vm10.local ceph-mon[51861]: pgmap v2292: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:43 vm08.local ceph-mon[47734]: pgmap v2292: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:09:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:09:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:09:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:09:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:45 vm10.local ceph-mon[51861]: pgmap v2293: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:45 vm08.local ceph-mon[47734]: pgmap v2293: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:46.907 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:46.907 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:46.938 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:46.938 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:47 vm10.local ceph-mon[51861]: pgmap v2294: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:47 vm08.local ceph-mon[47734]: pgmap v2294: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:49.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:49 vm08.local ceph-mon[47734]: pgmap v2295: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:49 vm10.local ceph-mon[51861]: pgmap v2295: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:51 vm10.local ceph-mon[51861]: pgmap v2296: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:51 vm08.local ceph-mon[47734]: pgmap v2296: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:51.939 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:51.940 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:51.967 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:51.968 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:53 vm10.local ceph-mon[51861]: pgmap v2297: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:53 vm08.local ceph-mon[47734]: pgmap v2297: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:55 vm10.local ceph-mon[51861]: pgmap v2298: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:55 vm08.local ceph-mon[47734]: pgmap v2298: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:09:56.969 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:09:56.969 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:09:56.997 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:09:56.998 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:09:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:57 vm10.local ceph-mon[51861]: pgmap v2299: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:57 vm08.local ceph-mon[47734]: pgmap v2299: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:59.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:09:59 vm08.local ceph-mon[47734]: pgmap v2300: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:09:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:09:59 vm10.local ceph-mon[51861]: pgmap v2300: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:00.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T15:10:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T15:10:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:01 vm08.local ceph-mon[47734]: pgmap v2301: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:01 vm10.local ceph-mon[51861]: pgmap v2301: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:01.999 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:02.000 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:02.026 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:02.027 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:02 vm08.local ceph-mon[47734]: pgmap v2302: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:02 vm10.local ceph-mon[51861]: pgmap v2302: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:05.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:05 vm08.local ceph-mon[47734]: pgmap v2303: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:05 vm10.local ceph-mon[51861]: pgmap v2303: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:07.028 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:07.029 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:07.054 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:07.054 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:07 vm08.local ceph-mon[47734]: pgmap v2304: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:07 vm10.local ceph-mon[51861]: pgmap v2304: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:09 vm08.local ceph-mon[47734]: pgmap v2305: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:09 vm10.local ceph-mon[51861]: pgmap v2305: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:11 vm08.local ceph-mon[47734]: pgmap v2306: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:11 vm10.local ceph-mon[51861]: pgmap v2306: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:12.055 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:12.056 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:12.099 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:12.100 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:13 vm08.local ceph-mon[47734]: pgmap v2307: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:13 vm10.local ceph-mon[51861]: pgmap v2307: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:15 vm10.local ceph-mon[51861]: pgmap v2308: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:15 vm08.local ceph-mon[47734]: pgmap v2308: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:17.101 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:17.101 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:17.126 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:17.126 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:17 vm10.local ceph-mon[51861]: pgmap v2309: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:17 vm08.local ceph-mon[47734]: pgmap v2309: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:19 vm08.local ceph-mon[47734]: pgmap v2310: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:10:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:10:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:10:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:10:19.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:10:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:19 vm10.local ceph-mon[51861]: pgmap v2310: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:10:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:10:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:10:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:10:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:10:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:21 vm10.local ceph-mon[51861]: pgmap v2311: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:21 vm08.local ceph-mon[47734]: pgmap v2311: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:22.128 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:22.128 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:22.154 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:22.154 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:23 vm10.local ceph-mon[51861]: pgmap v2312: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:23 vm08.local ceph-mon[47734]: pgmap v2312: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:25 vm10.local ceph-mon[51861]: pgmap v2313: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:25 vm08.local ceph-mon[47734]: pgmap v2313: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:27.156 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:27.156 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:27.183 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:27.183 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:27 vm10.local ceph-mon[51861]: pgmap v2314: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:27 vm08.local ceph-mon[47734]: pgmap v2314: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:29 vm08.local ceph-mon[47734]: pgmap v2315: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:29 vm10.local ceph-mon[51861]: pgmap v2315: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:31 vm10.local ceph-mon[51861]: pgmap v2316: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:31 vm08.local ceph-mon[47734]: pgmap v2316: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:32.185 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:32.185 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:32.212 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:32.213 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:33 vm10.local ceph-mon[51861]: pgmap v2317: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:33 vm08.local ceph-mon[47734]: pgmap v2317: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:35 vm10.local ceph-mon[51861]: pgmap v2318: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:35 vm08.local ceph-mon[47734]: pgmap v2318: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:37.214 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:37.214 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:37.244 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:37.245 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:37 vm10.local ceph-mon[51861]: pgmap v2319: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:37 vm08.local ceph-mon[47734]: pgmap v2319: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:39 vm08.local ceph-mon[47734]: pgmap v2320: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:39 vm10.local ceph-mon[51861]: pgmap v2320: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:41 vm10.local ceph-mon[51861]: pgmap v2321: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:41 vm08.local ceph-mon[47734]: pgmap v2321: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:42.246 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:42.247 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:42.274 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:42.274 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:43 vm10.local ceph-mon[51861]: pgmap v2322: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:10:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:10:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:43 vm08.local ceph-mon[47734]: pgmap v2322: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:10:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:10:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:45 vm10.local ceph-mon[51861]: pgmap v2323: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:45 vm08.local ceph-mon[47734]: pgmap v2323: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:47.276 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:47.276 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:47.303 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:47.304 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:47 vm10.local ceph-mon[51861]: pgmap v2324: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:47 vm08.local ceph-mon[47734]: pgmap v2324: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:49 vm08.local ceph-mon[47734]: pgmap v2325: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:49 vm10.local ceph-mon[51861]: pgmap v2325: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:51 vm08.local ceph-mon[47734]: pgmap v2326: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:51 vm10.local ceph-mon[51861]: pgmap v2326: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:52.305 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:52.306 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:52.333 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:52.334 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:53 vm08.local ceph-mon[47734]: pgmap v2327: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:53 vm10.local ceph-mon[51861]: pgmap v2327: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:55 vm08.local ceph-mon[47734]: pgmap v2328: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:55 vm10.local ceph-mon[51861]: pgmap v2328: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:10:57.335 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:10:57.335 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:10:57.360 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:10:57.361 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:10:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:57 vm08.local ceph-mon[47734]: pgmap v2329: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:57 vm10.local ceph-mon[51861]: pgmap v2329: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:10:59 vm08.local ceph-mon[47734]: pgmap v2330: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:10:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:10:59 vm10.local ceph-mon[51861]: pgmap v2330: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:01 vm10.local ceph-mon[51861]: pgmap v2331: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:01 vm08.local ceph-mon[47734]: pgmap v2331: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:02.362 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:02.362 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:02.388 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:02.388 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:03 vm10.local ceph-mon[51861]: pgmap v2332: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:03 vm08.local ceph-mon[47734]: pgmap v2332: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:05 vm10.local ceph-mon[51861]: pgmap v2333: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:05 vm08.local ceph-mon[47734]: pgmap v2333: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:07.389 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:07.390 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:07 vm10.local ceph-mon[51861]: pgmap v2334: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:07.417 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:07.417 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:07 vm08.local ceph-mon[47734]: pgmap v2334: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:09 vm08.local ceph-mon[47734]: pgmap v2335: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:09 vm10.local ceph-mon[51861]: pgmap v2335: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:11 vm10.local ceph-mon[51861]: pgmap v2336: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:11 vm08.local ceph-mon[47734]: pgmap v2336: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:12.419 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:12.419 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:12.444 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:12.445 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:13 vm10.local ceph-mon[51861]: pgmap v2337: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:13 vm08.local ceph-mon[47734]: pgmap v2337: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:15 vm10.local ceph-mon[51861]: pgmap v2338: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:15 vm08.local ceph-mon[47734]: pgmap v2338: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:17 vm10.local ceph-mon[51861]: pgmap v2339: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:17.446 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:17.447 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:17.472 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:17.473 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:17 vm08.local ceph-mon[47734]: pgmap v2339: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:19 vm10.local ceph-mon[51861]: pgmap v2340: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:11:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:11:19.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:11:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:19 vm08.local ceph-mon[47734]: pgmap v2340: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:11:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:11:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:11:20.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:11:20.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:11:20.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:11:20.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:11:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:21 vm10.local ceph-mon[51861]: pgmap v2341: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:21 vm08.local ceph-mon[47734]: pgmap v2341: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:22.474 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:22.475 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:22.501 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:22.501 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:23 vm10.local ceph-mon[51861]: pgmap v2342: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:23 vm08.local ceph-mon[47734]: pgmap v2342: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:24 vm08.local ceph-mon[47734]: pgmap v2343: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:24 vm10.local ceph-mon[51861]: pgmap v2343: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:26 vm10.local ceph-mon[51861]: pgmap v2344: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:26 vm08.local ceph-mon[47734]: pgmap v2344: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:27.502 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:27.503 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:27.534 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:27.535 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:29 vm08.local ceph-mon[47734]: pgmap v2345: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:29 vm10.local ceph-mon[51861]: pgmap v2345: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:31 vm10.local ceph-mon[51861]: pgmap v2346: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:31 vm08.local ceph-mon[47734]: pgmap v2346: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:32.536 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:32.536 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:32.744 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:32.745 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:33.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:32 vm08.local ceph-mon[47734]: pgmap v2347: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:33.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:32 vm10.local ceph-mon[51861]: pgmap v2347: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:35 vm10.local ceph-mon[51861]: pgmap v2348: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:35 vm08.local ceph-mon[47734]: pgmap v2348: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:36 vm10.local ceph-mon[51861]: pgmap v2349: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:36 vm08.local ceph-mon[47734]: pgmap v2349: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:37.746 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:37.747 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:37.773 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:37.774 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:39 vm08.local ceph-mon[47734]: pgmap v2350: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:39 vm10.local ceph-mon[51861]: pgmap v2350: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:41 vm10.local ceph-mon[51861]: pgmap v2351: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:41 vm08.local ceph-mon[47734]: pgmap v2351: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:42.775 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:42.775 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:42.801 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:42.802 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:43 vm10.local ceph-mon[51861]: pgmap v2352: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:11:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:11:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:43 vm08.local ceph-mon[47734]: pgmap v2352: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:11:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:11:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:45 vm10.local ceph-mon[51861]: pgmap v2353: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:45 vm08.local ceph-mon[47734]: pgmap v2353: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:47 vm10.local ceph-mon[51861]: pgmap v2354: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:47 vm08.local ceph-mon[47734]: pgmap v2354: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:47.803 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:47.804 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:47.829 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:47.830 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:49 vm08.local ceph-mon[47734]: pgmap v2355: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:49 vm10.local ceph-mon[51861]: pgmap v2355: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:50 vm08.local ceph-mon[47734]: pgmap v2356: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:50 vm10.local ceph-mon[51861]: pgmap v2356: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:52.831 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:52.832 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:52.860 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:52.860 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:53.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:52 vm10.local ceph-mon[51861]: pgmap v2357: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:52 vm08.local ceph-mon[47734]: pgmap v2357: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:55 vm10.local ceph-mon[51861]: pgmap v2358: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:55 vm08.local ceph-mon[47734]: pgmap v2358: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:11:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:57 vm10.local ceph-mon[51861]: pgmap v2359: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:57 vm08.local ceph-mon[47734]: pgmap v2359: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:57.861 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:11:57.862 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:11:57.903 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:11:57.903 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:11:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:11:59 vm08.local ceph-mon[47734]: pgmap v2360: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:11:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:11:59 vm10.local ceph-mon[51861]: pgmap v2360: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:01 vm08.local ceph-mon[47734]: pgmap v2361: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:01 vm10.local ceph-mon[51861]: pgmap v2361: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:02.904 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:02.905 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:02.970 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:02.971 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:03 vm10.local ceph-mon[51861]: pgmap v2362: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:03 vm08.local ceph-mon[47734]: pgmap v2362: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:05 vm08.local ceph-mon[47734]: pgmap v2363: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:05 vm10.local ceph-mon[51861]: pgmap v2363: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:07 vm08.local ceph-mon[47734]: pgmap v2364: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:07 vm10.local ceph-mon[51861]: pgmap v2364: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:07.972 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:07.973 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:08.002 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:08.002 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:09 vm08.local ceph-mon[47734]: pgmap v2365: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:09 vm10.local ceph-mon[51861]: pgmap v2365: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:11 vm08.local ceph-mon[47734]: pgmap v2366: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:11 vm10.local ceph-mon[51861]: pgmap v2366: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:13.004 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:13.004 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:13.031 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:13.031 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:13.460 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:13 vm10.local ceph-mon[51861]: pgmap v2367: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:13 vm08.local ceph-mon[47734]: pgmap v2367: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:14 vm08.local ceph-mon[47734]: pgmap v2368: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:14 vm10.local ceph-mon[51861]: pgmap v2368: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:16 vm10.local ceph-mon[51861]: pgmap v2369: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:16 vm08.local ceph-mon[47734]: pgmap v2369: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:18.033 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:18.033 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:18.058 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:18.059 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:19.274 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:19 vm10.local ceph-mon[51861]: pgmap v2370: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:19.274 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:12:19.274 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:12:19.274 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:19 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:12:19.278 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:19 vm08.local ceph-mon[47734]: pgmap v2370: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:19.278 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:12:19.278 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:12:19.278 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:19 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:12:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:12:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:12:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:20 vm08.local ceph-mon[47734]: pgmap v2371: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:12:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:12:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:20 vm10.local ceph-mon[51861]: pgmap v2371: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:23.060 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:23.060 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:23.087 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:23.088 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:22 vm10.local ceph-mon[51861]: pgmap v2372: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:22 vm08.local ceph-mon[47734]: pgmap v2372: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:25 vm10.local ceph-mon[51861]: pgmap v2373: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:25 vm08.local ceph-mon[47734]: pgmap v2373: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:27 vm10.local ceph-mon[51861]: pgmap v2374: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:27 vm08.local ceph-mon[47734]: pgmap v2374: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:28.089 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:28.089 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:28.118 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:28.119 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:29 vm08.local ceph-mon[47734]: pgmap v2375: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:29 vm10.local ceph-mon[51861]: pgmap v2375: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:31 vm08.local ceph-mon[47734]: pgmap v2376: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:31 vm10.local ceph-mon[51861]: pgmap v2376: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:32 vm08.local ceph-mon[47734]: pgmap v2377: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:32 vm10.local ceph-mon[51861]: pgmap v2377: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:33.121 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:33.121 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:33.306 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:33.306 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:35 vm08.local ceph-mon[47734]: pgmap v2378: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:35 vm10.local ceph-mon[51861]: pgmap v2378: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:36 vm08.local ceph-mon[47734]: pgmap v2379: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:36 vm10.local ceph-mon[51861]: pgmap v2379: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:38.308 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:38.308 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:38.435 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:38.435 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:39 vm08.local ceph-mon[47734]: pgmap v2380: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:39 vm10.local ceph-mon[51861]: pgmap v2380: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:41 vm08.local ceph-mon[47734]: pgmap v2381: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:41 vm10.local ceph-mon[51861]: pgmap v2381: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:42 vm08.local ceph-mon[47734]: pgmap v2382: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:42 vm10.local ceph-mon[51861]: pgmap v2382: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:43.437 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:43.437 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:43.464 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:43.465 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:12:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:12:43.559 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:12:43.559 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:12:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:44 vm08.local ceph-mon[47734]: pgmap v2383: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:44 vm10.local ceph-mon[51861]: pgmap v2383: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:46 vm08.local ceph-mon[47734]: pgmap v2384: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:46 vm10.local ceph-mon[51861]: pgmap v2384: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:48.466 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:48.467 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:48.493 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:48.493 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:49 vm08.local ceph-mon[47734]: pgmap v2385: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:49 vm10.local ceph-mon[51861]: pgmap v2385: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:51 vm08.local ceph-mon[47734]: pgmap v2386: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:51 vm10.local ceph-mon[51861]: pgmap v2386: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:53.495 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:53.495 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:53.522 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:53.523 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:53 vm08.local ceph-mon[47734]: pgmap v2387: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:53 vm10.local ceph-mon[51861]: pgmap v2387: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:55 vm08.local ceph-mon[47734]: pgmap v2388: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:55 vm10.local ceph-mon[51861]: pgmap v2388: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:12:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:57 vm08.local ceph-mon[47734]: pgmap v2389: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:57 vm10.local ceph-mon[51861]: pgmap v2389: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:58.525 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:12:58.525 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:12:58.578 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:12:58.578 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:12:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:12:59 vm08.local ceph-mon[47734]: pgmap v2390: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:12:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:12:59 vm10.local ceph-mon[51861]: pgmap v2390: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:00 vm08.local ceph-mon[47734]: pgmap v2391: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:00 vm10.local ceph-mon[51861]: pgmap v2391: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:02 vm08.local ceph-mon[47734]: pgmap v2392: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:02 vm10.local ceph-mon[51861]: pgmap v2392: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:03.580 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:03.581 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:03.608 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:03.608 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:05 vm08.local ceph-mon[47734]: pgmap v2393: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:05 vm10.local ceph-mon[51861]: pgmap v2393: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:07 vm08.local ceph-mon[47734]: pgmap v2394: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:07 vm10.local ceph-mon[51861]: pgmap v2394: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:08.609 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:08.610 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:08.664 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:08.664 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:09 vm08.local ceph-mon[47734]: pgmap v2395: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:09 vm10.local ceph-mon[51861]: pgmap v2395: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:11 vm08.local ceph-mon[47734]: pgmap v2396: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:11 vm10.local ceph-mon[51861]: pgmap v2396: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:13.460 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:13 vm10.local ceph-mon[51861]: pgmap v2397: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:13 vm08.local ceph-mon[47734]: pgmap v2397: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:13.665 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:13.666 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:13.697 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:13.698 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:14 vm08.local ceph-mon[47734]: pgmap v2398: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:14 vm10.local ceph-mon[51861]: pgmap v2398: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:16 vm08.local ceph-mon[47734]: pgmap v2399: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:13:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:16 vm10.local ceph-mon[51861]: pgmap v2399: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:13:18.699 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:18.699 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:18.725 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:18.725 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:19 vm08.local ceph-mon[47734]: pgmap v2400: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:13:19.315 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:19 vm10.local ceph-mon[51861]: pgmap v2400: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:13:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:13:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:13:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:13:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:13:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:13:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:13:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:13:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:21 vm08.local ceph-mon[47734]: pgmap v2401: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:21 vm10.local ceph-mon[51861]: pgmap v2401: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:23 vm08.local ceph-mon[47734]: pgmap v2402: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:13:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:23 vm10.local ceph-mon[51861]: pgmap v2402: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:13:23.726 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:23.727 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:23.753 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:23.754 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:25 vm08.local ceph-mon[47734]: pgmap v2403: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:25 vm10.local ceph-mon[51861]: pgmap v2403: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:27 vm08.local ceph-mon[47734]: pgmap v2404: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:27 vm10.local ceph-mon[51861]: pgmap v2404: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:28.755 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:28.755 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:28.781 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:28.781 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:29 vm08.local ceph-mon[47734]: pgmap v2405: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:29 vm10.local ceph-mon[51861]: pgmap v2405: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:31 vm08.local ceph-mon[47734]: pgmap v2406: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:31 vm10.local ceph-mon[51861]: pgmap v2406: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:33 vm08.local ceph-mon[47734]: pgmap v2407: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:33 vm10.local ceph-mon[51861]: pgmap v2407: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:33.783 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:33.783 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:33.809 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:33.810 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:35 vm08.local ceph-mon[47734]: pgmap v2408: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:35 vm10.local ceph-mon[51861]: pgmap v2408: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:37 vm08.local ceph-mon[47734]: pgmap v2409: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:37 vm10.local ceph-mon[51861]: pgmap v2409: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:38.811 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:38.812 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:38.839 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:38.839 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:39 vm08.local ceph-mon[47734]: pgmap v2410: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:39 vm10.local ceph-mon[51861]: pgmap v2410: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:41 vm08.local ceph-mon[47734]: pgmap v2411: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:41 vm10.local ceph-mon[51861]: pgmap v2411: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:42.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:42 vm08.local ceph-mon[47734]: pgmap v2412: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:42.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:42 vm10.local ceph-mon[51861]: pgmap v2412: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:13:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:13:43.559 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:13:43.559 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:13:43.841 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:43.841 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:43.872 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:43.873 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:44 vm08.local ceph-mon[47734]: pgmap v2413: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:44 vm10.local ceph-mon[51861]: pgmap v2413: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:46 vm08.local ceph-mon[47734]: pgmap v2414: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:46 vm10.local ceph-mon[51861]: pgmap v2414: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:48.874 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:48.874 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:48.904 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:48.905 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:49 vm08.local ceph-mon[47734]: pgmap v2415: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:49 vm10.local ceph-mon[51861]: pgmap v2415: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:51 vm08.local ceph-mon[47734]: pgmap v2416: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:51 vm10.local ceph-mon[51861]: pgmap v2416: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:53 vm08.local ceph-mon[47734]: pgmap v2417: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:53 vm10.local ceph-mon[51861]: pgmap v2417: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:53.906 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:53.906 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:53.941 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:53.941 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:55 vm08.local ceph-mon[47734]: pgmap v2418: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:55 vm10.local ceph-mon[51861]: pgmap v2418: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:13:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:56 vm08.local ceph-mon[47734]: pgmap v2419: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:56 vm10.local ceph-mon[51861]: pgmap v2419: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:58.943 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:13:58.943 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:13:58.973 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:13:58.974 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:13:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:13:59 vm08.local ceph-mon[47734]: pgmap v2420: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:13:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:13:59 vm10.local ceph-mon[51861]: pgmap v2420: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:01 vm08.local ceph-mon[47734]: pgmap v2421: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:01 vm10.local ceph-mon[51861]: pgmap v2421: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:03 vm08.local ceph-mon[47734]: pgmap v2422: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:03 vm10.local ceph-mon[51861]: pgmap v2422: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:03.975 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:03.975 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:04.001 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:04.001 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:04 vm08.local ceph-mon[47734]: pgmap v2423: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:04 vm10.local ceph-mon[51861]: pgmap v2423: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:06 vm08.local ceph-mon[47734]: pgmap v2424: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:06 vm10.local ceph-mon[51861]: pgmap v2424: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:09.003 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:09.003 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:09.029 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:09.029 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:09 vm08.local ceph-mon[47734]: pgmap v2425: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:09 vm10.local ceph-mon[51861]: pgmap v2425: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:11 vm08.local ceph-mon[47734]: pgmap v2426: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:11 vm10.local ceph-mon[51861]: pgmap v2426: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:13.460 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:13 vm10.local ceph-mon[51861]: pgmap v2427: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:13 vm08.local ceph-mon[47734]: pgmap v2427: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:14.031 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:14.031 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:14.057 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:14.058 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:15 vm08.local ceph-mon[47734]: pgmap v2428: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:15 vm10.local ceph-mon[51861]: pgmap v2428: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:16.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:16 vm08.local ceph-mon[47734]: pgmap v2429: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:16 vm10.local ceph-mon[51861]: pgmap v2429: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:19.059 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:19.060 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:19.086 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:19.087 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:19 vm08.local ceph-mon[47734]: pgmap v2430: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:19.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:19 vm10.local ceph-mon[51861]: pgmap v2430: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:14:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:14:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:14:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:14:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:14:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:14:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:14:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:14:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:14:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:21 vm08.local ceph-mon[47734]: pgmap v2431: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:21 vm10.local ceph-mon[51861]: pgmap v2431: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:23 vm08.local ceph-mon[47734]: pgmap v2432: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:23 vm10.local ceph-mon[51861]: pgmap v2432: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:24.088 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:24.089 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:24.115 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:24.115 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:25 vm08.local ceph-mon[47734]: pgmap v2433: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:25 vm10.local ceph-mon[51861]: pgmap v2433: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:26.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:26 vm08.local ceph-mon[47734]: pgmap v2434: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:26 vm10.local ceph-mon[51861]: pgmap v2434: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:29.117 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:29.117 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:29.144 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:29.145 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:29 vm08.local ceph-mon[47734]: pgmap v2435: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:29 vm10.local ceph-mon[51861]: pgmap v2435: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:31 vm08.local ceph-mon[47734]: pgmap v2436: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:31 vm10.local ceph-mon[51861]: pgmap v2436: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:33 vm08.local ceph-mon[47734]: pgmap v2437: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:33 vm10.local ceph-mon[51861]: pgmap v2437: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:34.146 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:34.147 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:34.205 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:34.205 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:34 vm08.local ceph-mon[47734]: pgmap v2438: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:34 vm10.local ceph-mon[51861]: pgmap v2438: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:37.067 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:36 vm08.local ceph-mon[47734]: pgmap v2439: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:36 vm10.local ceph-mon[51861]: pgmap v2439: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:39.207 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:39.207 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:39.234 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:39.234 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:39 vm08.local ceph-mon[47734]: pgmap v2440: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:39 vm10.local ceph-mon[51861]: pgmap v2440: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:41 vm08.local ceph-mon[47734]: pgmap v2441: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:41 vm10.local ceph-mon[51861]: pgmap v2441: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:43.460 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:43 vm10.local ceph-mon[51861]: pgmap v2442: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:43.460 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:14:43.460 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:14:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:43 vm08.local ceph-mon[47734]: pgmap v2442: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:14:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:14:44.236 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:44.236 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:44.263 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:44.264 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:45 vm08.local ceph-mon[47734]: pgmap v2443: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:45 vm10.local ceph-mon[51861]: pgmap v2443: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:46.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:46 vm08.local ceph-mon[47734]: pgmap v2444: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:46 vm10.local ceph-mon[51861]: pgmap v2444: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:49.265 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:49.265 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:49.293 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:49.293 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:49 vm08.local ceph-mon[47734]: pgmap v2445: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:49 vm10.local ceph-mon[51861]: pgmap v2445: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:51 vm08.local ceph-mon[47734]: pgmap v2446: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:51 vm10.local ceph-mon[51861]: pgmap v2446: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:53 vm08.local ceph-mon[47734]: pgmap v2447: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:53 vm10.local ceph-mon[51861]: pgmap v2447: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:54.294 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:54.294 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:54.320 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:54.321 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:54 vm08.local ceph-mon[47734]: pgmap v2448: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:54 vm10.local ceph-mon[51861]: pgmap v2448: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:14:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:56 vm08.local ceph-mon[47734]: pgmap v2449: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:56 vm10.local ceph-mon[51861]: pgmap v2449: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:14:59 vm08.local ceph-mon[47734]: pgmap v2450: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:14:59.322 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:14:59.322 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:14:59.348 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:14:59.349 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:14:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:14:59 vm10.local ceph-mon[51861]: pgmap v2450: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:01 vm08.local ceph-mon[47734]: pgmap v2451: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:01 vm10.local ceph-mon[51861]: pgmap v2451: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:02.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:02 vm08.local ceph-mon[47734]: pgmap v2452: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:02.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:02 vm10.local ceph-mon[51861]: pgmap v2452: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:04.350 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:04.351 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:04.377 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:04.377 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:05 vm08.local ceph-mon[47734]: pgmap v2453: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:05 vm10.local ceph-mon[51861]: pgmap v2453: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:07 vm08.local ceph-mon[47734]: pgmap v2454: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:07 vm10.local ceph-mon[51861]: pgmap v2454: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:09 vm08.local ceph-mon[47734]: pgmap v2455: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:09.379 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:09.379 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:09.406 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:09.407 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:09 vm10.local ceph-mon[51861]: pgmap v2455: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:11 vm08.local ceph-mon[47734]: pgmap v2456: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:11 vm10.local ceph-mon[51861]: pgmap v2456: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:12.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:12 vm08.local ceph-mon[47734]: pgmap v2457: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:12.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:12 vm10.local ceph-mon[51861]: pgmap v2457: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:14.408 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:14.409 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:14.434 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:14.435 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:15 vm08.local ceph-mon[47734]: pgmap v2458: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:15 vm10.local ceph-mon[51861]: pgmap v2458: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:17 vm08.local ceph-mon[47734]: pgmap v2459: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:17 vm10.local ceph-mon[51861]: pgmap v2459: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:19 vm08.local ceph-mon[47734]: pgmap v2460: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:19.436 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:19.437 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:19.462 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:19.463 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:19 vm10.local ceph-mon[51861]: pgmap v2460: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:20.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:20.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:20.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:20.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:15:20.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:15:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:15:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:15:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:15:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:15:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:15:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:21 vm08.local ceph-mon[47734]: pgmap v2461: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:21 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:15:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:21 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:15:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:21 vm10.local ceph-mon[51861]: pgmap v2461: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:21 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:15:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:21 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:15:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:22 vm08.local ceph-mon[47734]: pgmap v2462: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:22 vm10.local ceph-mon[51861]: pgmap v2462: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:24.464 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:24.465 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:24.492 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:24.493 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:25 vm08.local ceph-mon[47734]: pgmap v2463: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:25 vm10.local ceph-mon[51861]: pgmap v2463: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:27 vm08.local ceph-mon[47734]: pgmap v2464: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:27 vm10.local ceph-mon[51861]: pgmap v2464: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:29 vm08.local ceph-mon[47734]: pgmap v2465: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:29.494 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:29.494 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:29.523 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:29.524 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:29 vm10.local ceph-mon[51861]: pgmap v2465: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:31 vm08.local ceph-mon[47734]: pgmap v2466: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:31 vm10.local ceph-mon[51861]: pgmap v2466: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:32.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:32 vm08.local ceph-mon[47734]: pgmap v2467: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:32 vm10.local ceph-mon[51861]: pgmap v2467: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:34.525 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:34.525 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:34.552 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:34.553 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:35 vm08.local ceph-mon[47734]: pgmap v2468: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:35 vm10.local ceph-mon[51861]: pgmap v2468: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:36.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:36 vm08.local ceph-mon[47734]: pgmap v2469: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:36 vm10.local ceph-mon[51861]: pgmap v2469: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:39 vm08.local ceph-mon[47734]: pgmap v2470: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:39.554 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:39.555 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:39.584 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:39.584 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:39 vm10.local ceph-mon[51861]: pgmap v2470: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:40 vm08.local ceph-mon[47734]: pgmap v2471: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:40 vm10.local ceph-mon[51861]: pgmap v2471: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:42 vm08.local ceph-mon[47734]: pgmap v2472: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:42 vm10.local ceph-mon[51861]: pgmap v2472: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:15:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:15:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:15:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:15:44.586 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:44.586 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:44.611 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:44.611 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:45 vm08.local ceph-mon[47734]: pgmap v2473: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:45 vm10.local ceph-mon[51861]: pgmap v2473: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:46.478 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:46 vm08.local ceph-mon[47734]: pgmap v2474: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:46.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:46 vm10.local ceph-mon[51861]: pgmap v2474: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:49 vm08.local ceph-mon[47734]: pgmap v2475: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:49.613 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:49.613 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:49.640 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:49.640 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:49 vm10.local ceph-mon[51861]: pgmap v2475: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:50 vm08.local ceph-mon[47734]: pgmap v2476: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:50 vm10.local ceph-mon[51861]: pgmap v2476: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:52 vm08.local ceph-mon[47734]: pgmap v2477: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:52 vm10.local ceph-mon[51861]: pgmap v2477: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:54.642 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:54.642 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:54.668 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:54.668 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:15:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:55 vm08.local ceph-mon[47734]: pgmap v2478: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:55 vm10.local ceph-mon[51861]: pgmap v2478: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:15:56.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:56 vm08.local ceph-mon[47734]: pgmap v2479: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:56.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:56 vm10.local ceph-mon[51861]: pgmap v2479: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:15:59 vm08.local ceph-mon[47734]: pgmap v2480: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:15:59 vm10.local ceph-mon[51861]: pgmap v2480: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:15:59.669 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:15:59.670 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:15:59.696 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:15:59.696 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:00 vm08.local ceph-mon[47734]: pgmap v2481: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:00 vm10.local ceph-mon[51861]: pgmap v2481: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:02 vm08.local ceph-mon[47734]: pgmap v2482: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:02 vm10.local ceph-mon[51861]: pgmap v2482: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:04.697 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:04.698 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:04.723 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:04.724 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:05 vm08.local ceph-mon[47734]: pgmap v2483: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:05 vm10.local ceph-mon[51861]: pgmap v2483: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:06.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:06 vm08.local ceph-mon[47734]: pgmap v2484: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:06.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:06 vm10.local ceph-mon[51861]: pgmap v2484: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:09 vm08.local ceph-mon[47734]: pgmap v2485: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:09 vm10.local ceph-mon[51861]: pgmap v2485: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:09.725 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:09.725 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:09.752 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:09.753 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:10.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:10 vm08.local ceph-mon[47734]: pgmap v2486: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:10.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:10 vm10.local ceph-mon[51861]: pgmap v2486: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:12 vm08.local ceph-mon[47734]: pgmap v2487: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:12 vm10.local ceph-mon[51861]: pgmap v2487: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:14.754 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:14.754 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:14.782 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:14.783 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:15 vm08.local ceph-mon[47734]: pgmap v2488: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:15 vm10.local ceph-mon[51861]: pgmap v2488: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:16.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:16 vm08.local ceph-mon[47734]: pgmap v2489: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:16.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:16 vm10.local ceph-mon[51861]: pgmap v2489: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:19 vm08.local ceph-mon[47734]: pgmap v2490: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:19 vm10.local ceph-mon[51861]: pgmap v2490: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:19.784 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:19.785 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:19.811 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:19.811 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:20.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:20 vm08.local ceph-mon[47734]: pgmap v2491: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:20.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:20 vm10.local ceph-mon[51861]: pgmap v2491: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:21 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:16:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:21 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:16:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:21 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:16:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:21 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:16:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:21 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:16:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:21 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:16:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:21 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:16:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:21 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:16:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:21 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:16:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:21 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:16:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:21 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:16:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:21 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:16:22.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:22 vm08.local ceph-mon[47734]: pgmap v2492: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:22.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:22 vm10.local ceph-mon[51861]: pgmap v2492: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:24 vm08.local ceph-mon[47734]: pgmap v2493: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:24 vm10.local ceph-mon[51861]: pgmap v2493: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:24.812 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:24.813 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:24.839 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:24.840 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:26 vm08.local ceph-mon[47734]: pgmap v2494: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:26 vm10.local ceph-mon[51861]: pgmap v2494: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:29 vm08.local ceph-mon[47734]: pgmap v2495: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:29 vm10.local ceph-mon[51861]: pgmap v2495: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:29.841 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:29.842 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:29.867 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:29.867 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:30 vm08.local ceph-mon[47734]: pgmap v2496: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:30 vm10.local ceph-mon[51861]: pgmap v2496: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:32 vm08.local ceph-mon[47734]: pgmap v2497: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:32 vm10.local ceph-mon[51861]: pgmap v2497: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:34 vm08.local ceph-mon[47734]: pgmap v2498: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:34 vm10.local ceph-mon[51861]: pgmap v2498: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:34.868 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:34.869 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:34.895 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:34.895 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:36 vm08.local ceph-mon[47734]: pgmap v2499: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:36 vm10.local ceph-mon[51861]: pgmap v2499: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:39 vm08.local ceph-mon[47734]: pgmap v2500: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:39 vm10.local ceph-mon[51861]: pgmap v2500: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:39.896 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:39.897 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:39.924 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:39.924 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:40.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:40 vm08.local ceph-mon[47734]: pgmap v2501: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:40.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:40 vm10.local ceph-mon[51861]: pgmap v2501: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:42 vm08.local ceph-mon[47734]: pgmap v2502: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:42 vm10.local ceph-mon[51861]: pgmap v2502: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:16:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:16:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:16:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:16:44.925 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:44.926 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:44.953 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:44.953 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:44 vm08.local ceph-mon[47734]: pgmap v2503: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:45.309 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:44 vm10.local ceph-mon[51861]: pgmap v2503: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:46 vm08.local ceph-mon[47734]: pgmap v2504: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:46 vm10.local ceph-mon[51861]: pgmap v2504: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:49 vm08.local ceph-mon[47734]: pgmap v2505: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:49 vm10.local ceph-mon[51861]: pgmap v2505: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:49.955 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:49.955 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:49.981 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:49.981 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:50.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:50 vm08.local ceph-mon[47734]: pgmap v2506: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:50.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:50 vm10.local ceph-mon[51861]: pgmap v2506: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:52 vm08.local ceph-mon[47734]: pgmap v2507: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:52 vm10.local ceph-mon[51861]: pgmap v2507: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:54 vm08.local ceph-mon[47734]: pgmap v2508: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:54 vm10.local ceph-mon[51861]: pgmap v2508: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:16:54.983 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:16:54.983 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:16:55.009 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:16:55.010 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:16:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:56 vm08.local ceph-mon[47734]: pgmap v2509: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:56 vm10.local ceph-mon[51861]: pgmap v2509: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:16:59 vm08.local ceph-mon[47734]: pgmap v2510: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:16:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:16:59 vm10.local ceph-mon[51861]: pgmap v2510: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:00.011 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:00.012 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:00.037 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:00.037 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:00 vm08.local ceph-mon[47734]: pgmap v2511: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:00.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:00 vm10.local ceph-mon[51861]: pgmap v2511: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:02 vm08.local ceph-mon[47734]: pgmap v2512: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:02 vm10.local ceph-mon[51861]: pgmap v2512: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:04 vm08.local ceph-mon[47734]: pgmap v2513: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:04 vm10.local ceph-mon[51861]: pgmap v2513: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:05.038 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:05.039 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:05.064 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:05.064 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:06 vm08.local ceph-mon[47734]: pgmap v2514: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:06 vm10.local ceph-mon[51861]: pgmap v2514: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:08 vm08.local ceph-mon[47734]: pgmap v2515: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:08 vm10.local ceph-mon[51861]: pgmap v2515: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:10.066 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:10.066 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:10.094 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:10.094 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:10 vm08.local ceph-mon[47734]: pgmap v2516: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:10 vm10.local ceph-mon[51861]: pgmap v2516: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:13 vm08.local ceph-mon[47734]: pgmap v2517: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:13 vm10.local ceph-mon[51861]: pgmap v2517: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:14 vm08.local ceph-mon[47734]: pgmap v2518: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:14 vm10.local ceph-mon[51861]: pgmap v2518: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:15.095 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:15.096 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:15.121 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:15.121 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:16 vm08.local ceph-mon[47734]: pgmap v2519: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:17.361 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:16 vm10.local ceph-mon[51861]: pgmap v2519: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:18 vm08.local ceph-mon[47734]: pgmap v2520: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:18 vm10.local ceph-mon[51861]: pgmap v2520: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:20.122 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:20.123 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:20.149 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:20.149 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:20 vm08.local ceph-mon[47734]: pgmap v2521: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:17:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:17:21.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:20 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:17:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:20 vm10.local ceph-mon[51861]: pgmap v2521: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:17:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:17:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:20 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:17:22.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:22 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:17:22.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:22 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:17:22.393 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:22 vm08.local ceph-mon[47734]: pgmap v2522: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:22 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:17:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:22 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:17:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:22 vm10.local ceph-mon[51861]: pgmap v2522: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:17:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:17:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:17:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:17:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:17:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:17:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:24 vm10.local ceph-mon[51861]: pgmap v2523: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:24 vm08.local ceph-mon[47734]: pgmap v2523: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:25.151 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:25.151 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:25.177 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:25.178 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:26 vm08.local ceph-mon[47734]: pgmap v2524: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:26 vm10.local ceph-mon[51861]: pgmap v2524: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:29 vm08.local ceph-mon[47734]: pgmap v2525: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:29 vm10.local ceph-mon[51861]: pgmap v2525: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:30.179 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:30.180 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:30.206 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:30.206 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:31 vm08.local ceph-mon[47734]: pgmap v2526: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:31 vm10.local ceph-mon[51861]: pgmap v2526: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:33 vm08.local ceph-mon[47734]: pgmap v2527: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:33 vm10.local ceph-mon[51861]: pgmap v2527: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:34.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:34 vm08.local ceph-mon[47734]: pgmap v2528: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:34 vm10.local ceph-mon[51861]: pgmap v2528: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:35.208 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:35.208 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:35.235 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:35.235 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:37.061 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:36 vm08.local ceph-mon[47734]: pgmap v2529: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:36 vm10.local ceph-mon[51861]: pgmap v2529: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:38 vm08.local ceph-mon[47734]: pgmap v2530: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:38 vm10.local ceph-mon[51861]: pgmap v2530: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:40.237 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:40.238 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:40.265 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:40.265 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:40 vm08.local ceph-mon[47734]: pgmap v2531: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:40 vm10.local ceph-mon[51861]: pgmap v2531: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:42 vm08.local ceph-mon[47734]: pgmap v2532: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:42 vm10.local ceph-mon[51861]: pgmap v2532: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:17:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:17:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:17:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:17:45.267 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:45.267 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:45.296 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:45.297 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:44 vm08.local ceph-mon[47734]: pgmap v2533: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:44 vm10.local ceph-mon[51861]: pgmap v2533: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:47 vm08.local ceph-mon[47734]: pgmap v2534: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:47.348 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:47 vm10.local ceph-mon[51861]: pgmap v2534: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:49 vm08.local ceph-mon[47734]: pgmap v2535: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:49 vm10.local ceph-mon[51861]: pgmap v2535: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:50.299 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:50.299 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:50.324 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:50.325 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:51 vm08.local ceph-mon[47734]: pgmap v2536: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:51 vm10.local ceph-mon[51861]: pgmap v2536: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:53 vm08.local ceph-mon[47734]: pgmap v2537: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:53 vm10.local ceph-mon[51861]: pgmap v2537: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:54 vm08.local ceph-mon[47734]: pgmap v2538: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:54 vm10.local ceph-mon[51861]: pgmap v2538: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:17:55.326 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:17:55.326 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:17:55.352 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:17:55.353 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:17:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:56 vm08.local ceph-mon[47734]: pgmap v2539: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:56 vm10.local ceph-mon[51861]: pgmap v2539: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:17:59 vm08.local ceph-mon[47734]: pgmap v2540: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:17:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:17:59 vm10.local ceph-mon[51861]: pgmap v2540: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:00.354 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:00.354 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:00.380 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:00.380 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:01 vm08.local ceph-mon[47734]: pgmap v2541: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:01 vm10.local ceph-mon[51861]: pgmap v2541: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:03 vm08.local ceph-mon[47734]: pgmap v2542: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:03 vm10.local ceph-mon[51861]: pgmap v2542: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:04.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:04 vm08.local ceph-mon[47734]: pgmap v2543: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:04 vm10.local ceph-mon[51861]: pgmap v2543: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:05.382 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:05.382 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:05.408 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:05.408 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:06 vm08.local ceph-mon[47734]: pgmap v2544: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:06 vm10.local ceph-mon[51861]: pgmap v2544: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:08 vm08.local ceph-mon[47734]: pgmap v2545: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:08 vm10.local ceph-mon[51861]: pgmap v2545: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:10.409 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:10.409 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:10.435 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:10.435 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:10 vm08.local ceph-mon[47734]: pgmap v2546: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:10 vm10.local ceph-mon[51861]: pgmap v2546: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:13 vm08.local ceph-mon[47734]: pgmap v2547: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:13 vm10.local ceph-mon[51861]: pgmap v2547: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:14.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:14 vm08.local ceph-mon[47734]: pgmap v2548: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:14 vm10.local ceph-mon[51861]: pgmap v2548: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:15.437 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:15.437 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:15.463 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:15.463 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:16 vm08.local ceph-mon[47734]: pgmap v2549: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:17.360 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:16 vm10.local ceph-mon[51861]: pgmap v2549: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:18 vm08.local ceph-mon[47734]: pgmap v2550: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:18 vm10.local ceph-mon[51861]: pgmap v2550: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:20.464 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:20.465 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:20.511 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:20.511 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:20 vm08.local ceph-mon[47734]: pgmap v2551: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:20 vm10.local ceph-mon[51861]: pgmap v2551: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:22 vm08.local ceph-mon[47734]: pgmap v2552: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:22 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:18:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:22 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:18:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:22 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:18:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:22 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:18:23.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:22 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:18:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:22 vm10.local ceph-mon[51861]: pgmap v2552: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:22 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:18:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:22 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:18:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:22 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:18:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:22 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:18:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:22 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:18:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:24 vm08.local ceph-mon[47734]: pgmap v2553: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:24 vm10.local ceph-mon[51861]: pgmap v2553: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:25.513 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:25.514 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:25.542 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:25.544 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:26 vm08.local ceph-mon[47734]: pgmap v2554: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:26 vm10.local ceph-mon[51861]: pgmap v2554: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:28 vm08.local ceph-mon[47734]: pgmap v2555: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:28 vm10.local ceph-mon[51861]: pgmap v2555: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:30.545 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:30.546 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:30.572 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:30.572 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:30 vm08.local ceph-mon[47734]: pgmap v2556: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:30 vm10.local ceph-mon[51861]: pgmap v2556: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:32 vm08.local ceph-mon[47734]: pgmap v2557: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:32 vm10.local ceph-mon[51861]: pgmap v2557: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:34 vm08.local ceph-mon[47734]: pgmap v2558: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:34 vm10.local ceph-mon[51861]: pgmap v2558: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:35.574 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:35.574 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:35.602 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:35.602 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:36 vm08.local ceph-mon[47734]: pgmap v2559: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:36 vm10.local ceph-mon[51861]: pgmap v2559: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:38 vm08.local ceph-mon[47734]: pgmap v2560: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:38 vm10.local ceph-mon[51861]: pgmap v2560: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:40.603 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:40.604 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:40.630 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:40.630 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:40 vm08.local ceph-mon[47734]: pgmap v2561: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:40 vm10.local ceph-mon[51861]: pgmap v2561: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:42 vm08.local ceph-mon[47734]: pgmap v2562: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:42 vm10.local ceph-mon[51861]: pgmap v2562: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:18:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:18:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:18:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:18:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:45 vm08.local ceph-mon[47734]: pgmap v2563: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:45 vm10.local ceph-mon[51861]: pgmap v2563: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:45.631 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:45.632 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:45.658 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:45.658 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:47 vm08.local ceph-mon[47734]: pgmap v2564: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:47 vm10.local ceph-mon[51861]: pgmap v2564: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:49 vm08.local ceph-mon[47734]: pgmap v2565: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:49 vm10.local ceph-mon[51861]: pgmap v2565: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:50.660 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:50.660 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:50.686 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:50.687 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:51 vm08.local ceph-mon[47734]: pgmap v2566: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:51 vm10.local ceph-mon[51861]: pgmap v2566: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:53.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:53 vm08.local ceph-mon[47734]: pgmap v2567: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:53 vm10.local ceph-mon[51861]: pgmap v2567: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:54.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:54 vm08.local ceph-mon[47734]: pgmap v2568: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:54.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:54 vm10.local ceph-mon[51861]: pgmap v2568: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:18:55.688 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:18:55.689 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:18:55.715 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:18:55.715 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:18:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:56 vm08.local ceph-mon[47734]: pgmap v2569: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:56 vm10.local ceph-mon[51861]: pgmap v2569: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:18:58 vm08.local ceph-mon[47734]: pgmap v2570: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:18:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:18:58 vm10.local ceph-mon[51861]: pgmap v2570: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:00.716 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:00.717 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:00.851 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:00.851 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:01 vm10.local ceph-mon[51861]: pgmap v2571: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:01 vm08.local ceph-mon[47734]: pgmap v2571: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:03 vm10.local ceph-mon[51861]: pgmap v2572: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:03 vm08.local ceph-mon[47734]: pgmap v2572: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:04.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:04 vm10.local ceph-mon[51861]: pgmap v2573: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:04.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:04 vm08.local ceph-mon[47734]: pgmap v2573: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:05.853 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:05.853 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:05.880 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:05.881 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:06 vm08.local ceph-mon[47734]: pgmap v2574: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:06 vm10.local ceph-mon[51861]: pgmap v2574: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:09 vm08.local ceph-mon[47734]: pgmap v2575: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:09 vm10.local ceph-mon[51861]: pgmap v2575: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:10.882 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:10.883 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:10.908 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:10.909 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:11 vm08.local ceph-mon[47734]: pgmap v2576: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:11 vm10.local ceph-mon[51861]: pgmap v2576: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:13 vm08.local ceph-mon[47734]: pgmap v2577: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:13 vm10.local ceph-mon[51861]: pgmap v2577: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:14.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:14 vm10.local ceph-mon[51861]: pgmap v2578: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:14 vm08.local ceph-mon[47734]: pgmap v2578: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:15.910 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:15.911 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:15.937 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:15.937 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:16 vm08.local ceph-mon[47734]: pgmap v2579: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:17.361 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:16 vm10.local ceph-mon[51861]: pgmap v2579: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:18 vm08.local ceph-mon[47734]: pgmap v2580: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:18 vm10.local ceph-mon[51861]: pgmap v2580: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:20.939 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:20.939 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:20.969 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:20.970 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:21 vm08.local ceph-mon[47734]: pgmap v2581: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:21 vm10.local ceph-mon[51861]: pgmap v2581: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:23 vm08.local ceph-mon[47734]: pgmap v2582: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:19:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:19:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:19:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:19:23.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:19:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:23 vm10.local ceph-mon[51861]: pgmap v2582: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:19:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:19:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:19:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:19:23.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:19:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:24 vm10.local ceph-mon[51861]: pgmap v2583: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:24 vm08.local ceph-mon[47734]: pgmap v2583: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:25.971 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:25.972 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:25.998 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:25.998 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:26 vm08.local ceph-mon[47734]: pgmap v2584: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:26 vm10.local ceph-mon[51861]: pgmap v2584: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:29 vm08.local ceph-mon[47734]: pgmap v2585: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:29 vm10.local ceph-mon[51861]: pgmap v2585: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:30.999 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:31.000 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:31.025 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:31.025 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:31 vm08.local ceph-mon[47734]: pgmap v2586: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:31 vm10.local ceph-mon[51861]: pgmap v2586: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:33 vm08.local ceph-mon[47734]: pgmap v2587: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:33 vm10.local ceph-mon[51861]: pgmap v2587: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:34 vm10.local ceph-mon[51861]: pgmap v2588: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:34.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:34 vm08.local ceph-mon[47734]: pgmap v2588: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:36.027 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:36.027 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:36.055 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:36.055 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:36 vm08.local ceph-mon[47734]: pgmap v2589: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:36 vm10.local ceph-mon[51861]: pgmap v2589: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:39 vm08.local ceph-mon[47734]: pgmap v2590: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:39 vm10.local ceph-mon[51861]: pgmap v2590: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:41.057 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:41.057 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:41.083 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:41.083 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:41 vm08.local ceph-mon[47734]: pgmap v2591: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:41 vm10.local ceph-mon[51861]: pgmap v2591: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:43 vm08.local ceph-mon[47734]: pgmap v2592: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:43 vm10.local ceph-mon[51861]: pgmap v2592: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:19:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:19:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:19:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:19:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:45 vm08.local ceph-mon[47734]: pgmap v2593: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:45 vm10.local ceph-mon[51861]: pgmap v2593: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:46.085 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:46.085 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:46.111 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:46.111 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:47 vm08.local ceph-mon[47734]: pgmap v2594: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:47 vm10.local ceph-mon[51861]: pgmap v2594: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:49 vm08.local ceph-mon[47734]: pgmap v2595: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:49 vm10.local ceph-mon[51861]: pgmap v2595: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:51.113 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:51.113 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:51.139 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:51.139 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:51 vm08.local ceph-mon[47734]: pgmap v2596: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:51 vm10.local ceph-mon[51861]: pgmap v2596: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:53 vm10.local ceph-mon[51861]: pgmap v2597: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:53 vm08.local ceph-mon[47734]: pgmap v2597: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:54 vm08.local ceph-mon[47734]: pgmap v2598: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:54 vm10.local ceph-mon[51861]: pgmap v2598: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:19:56.140 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:19:56.141 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:19:56.167 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:19:56.168 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:19:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:57 vm08.local ceph-mon[47734]: pgmap v2599: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:57 vm10.local ceph-mon[51861]: pgmap v2599: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:19:59 vm08.local ceph-mon[47734]: pgmap v2600: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:19:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:19:59 vm10.local ceph-mon[51861]: pgmap v2600: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:00.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T15:20:00.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T15:20:01.169 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:01.169 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:01.196 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:01.197 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:01 vm08.local ceph-mon[47734]: pgmap v2601: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:01 vm10.local ceph-mon[51861]: pgmap v2601: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:03 vm08.local ceph-mon[47734]: pgmap v2602: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:03 vm10.local ceph-mon[51861]: pgmap v2602: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:04.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:04 vm08.local ceph-mon[47734]: pgmap v2603: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:04 vm10.local ceph-mon[51861]: pgmap v2603: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:06.198 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:06.199 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:06.226 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:06.226 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:07 vm08.local ceph-mon[47734]: pgmap v2604: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:07 vm10.local ceph-mon[51861]: pgmap v2604: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:09 vm08.local ceph-mon[47734]: pgmap v2605: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:09 vm10.local ceph-mon[51861]: pgmap v2605: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:11.228 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:11.228 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:11.254 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:11.255 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:11 vm08.local ceph-mon[47734]: pgmap v2606: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:11 vm10.local ceph-mon[51861]: pgmap v2606: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:13 vm08.local ceph-mon[47734]: pgmap v2607: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:13 vm10.local ceph-mon[51861]: pgmap v2607: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:14 vm08.local ceph-mon[47734]: pgmap v2608: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:14 vm10.local ceph-mon[51861]: pgmap v2608: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:16.256 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:16.257 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:16.282 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:16.282 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:17 vm08.local ceph-mon[47734]: pgmap v2609: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:17 vm10.local ceph-mon[51861]: pgmap v2609: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:19 vm08.local ceph-mon[47734]: pgmap v2610: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:19 vm10.local ceph-mon[51861]: pgmap v2610: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:21.284 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:21.284 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:21 vm08.local ceph-mon[47734]: pgmap v2611: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:21.309 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:21.309 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:21 vm10.local ceph-mon[51861]: pgmap v2611: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:23 vm08.local ceph-mon[47734]: pgmap v2612: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:20:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:20:23.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:23 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:20:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:23 vm10.local ceph-mon[51861]: pgmap v2612: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:20:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:20:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:23 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:20:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:20:24.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:20:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:20:24.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:20:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:25 vm08.local ceph-mon[47734]: pgmap v2613: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:25 vm10.local ceph-mon[51861]: pgmap v2613: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:26.311 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:26.311 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:26.337 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:26.338 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:26.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:26 vm10.local ceph-mon[51861]: pgmap v2614: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:26 vm08.local ceph-mon[47734]: pgmap v2614: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:29 vm08.local ceph-mon[47734]: pgmap v2615: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:29 vm10.local ceph-mon[51861]: pgmap v2615: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:31 vm08.local ceph-mon[47734]: pgmap v2616: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:31.339 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:31.340 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:31.366 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:31.367 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:31 vm10.local ceph-mon[51861]: pgmap v2616: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:33 vm08.local ceph-mon[47734]: pgmap v2617: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:33 vm10.local ceph-mon[51861]: pgmap v2617: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:34.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:34 vm08.local ceph-mon[47734]: pgmap v2618: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:34 vm10.local ceph-mon[51861]: pgmap v2618: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:36.368 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:36.369 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:36.395 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:36.395 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:37 vm08.local ceph-mon[47734]: pgmap v2619: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:37 vm10.local ceph-mon[51861]: pgmap v2619: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:39 vm08.local ceph-mon[47734]: pgmap v2620: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:39 vm10.local ceph-mon[51861]: pgmap v2620: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:41 vm08.local ceph-mon[47734]: pgmap v2621: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:41.397 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:41.398 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:41 vm10.local ceph-mon[51861]: pgmap v2621: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:41.425 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:41.425 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:43.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:43 vm08.local ceph-mon[47734]: pgmap v2622: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:43 vm10.local ceph-mon[51861]: pgmap v2622: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:20:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:20:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:20:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:20:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:45 vm08.local ceph-mon[47734]: pgmap v2623: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:45 vm10.local ceph-mon[51861]: pgmap v2623: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:46.427 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:46.427 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:46.453 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:46.454 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:47.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:47 vm08.local ceph-mon[47734]: pgmap v2624: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:47 vm10.local ceph-mon[51861]: pgmap v2624: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:49.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:49 vm08.local ceph-mon[47734]: pgmap v2625: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:49 vm10.local ceph-mon[51861]: pgmap v2625: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:51 vm08.local ceph-mon[47734]: pgmap v2626: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:51 vm10.local ceph-mon[51861]: pgmap v2626: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:51.455 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:51.456 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:51.484 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:51.484 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:53 vm10.local ceph-mon[51861]: pgmap v2627: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:53 vm08.local ceph-mon[47734]: pgmap v2627: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:54 vm08.local ceph-mon[47734]: pgmap v2628: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:54 vm10.local ceph-mon[51861]: pgmap v2628: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:20:56.486 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:20:56.487 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:20:56.513 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:20:56.514 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:20:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:57 vm08.local ceph-mon[47734]: pgmap v2629: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:57 vm10.local ceph-mon[51861]: pgmap v2629: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:20:59 vm08.local ceph-mon[47734]: pgmap v2630: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:20:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:20:59 vm10.local ceph-mon[51861]: pgmap v2630: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:01 vm08.local ceph-mon[47734]: pgmap v2631: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:01 vm10.local ceph-mon[51861]: pgmap v2631: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:01.515 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:01.516 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:01.544 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:01.544 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:03 vm08.local ceph-mon[47734]: pgmap v2632: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:03 vm10.local ceph-mon[51861]: pgmap v2632: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:04 vm08.local ceph-mon[47734]: pgmap v2633: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:04 vm10.local ceph-mon[51861]: pgmap v2633: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:06.546 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:06.546 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:06.572 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:06.573 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:07 vm08.local ceph-mon[47734]: pgmap v2634: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:07 vm10.local ceph-mon[51861]: pgmap v2634: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:09 vm08.local ceph-mon[47734]: pgmap v2635: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:09 vm10.local ceph-mon[51861]: pgmap v2635: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:11 vm08.local ceph-mon[47734]: pgmap v2636: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:11 vm10.local ceph-mon[51861]: pgmap v2636: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:11.574 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:11.574 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:11.601 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:11.601 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:13.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:13 vm08.local ceph-mon[47734]: pgmap v2637: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:13 vm10.local ceph-mon[51861]: pgmap v2637: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:14.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:14 vm08.local ceph-mon[47734]: pgmap v2638: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:14 vm10.local ceph-mon[51861]: pgmap v2638: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:16.603 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:16.603 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:16.630 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:16.631 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:17 vm08.local ceph-mon[47734]: pgmap v2639: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:17 vm10.local ceph-mon[51861]: pgmap v2639: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:19 vm08.local ceph-mon[47734]: pgmap v2640: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:19 vm10.local ceph-mon[51861]: pgmap v2640: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:21 vm10.local ceph-mon[51861]: pgmap v2641: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:21 vm08.local ceph-mon[47734]: pgmap v2641: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:21.632 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:21.633 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:21.660 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:21.660 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:23 vm10.local ceph-mon[51861]: pgmap v2642: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:23.494 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:23 vm08.local ceph-mon[47734]: pgmap v2642: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:24.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:21:24.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:21:24.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:21:24.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:21:24.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:21:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:21:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:21:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:21:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:21:24.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:21:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:25 vm10.local ceph-mon[51861]: pgmap v2643: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:25 vm08.local ceph-mon[47734]: pgmap v2643: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:26.662 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:26.662 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:26.690 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:26.691 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:27 vm10.local ceph-mon[51861]: pgmap v2644: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:27 vm08.local ceph-mon[47734]: pgmap v2644: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:29 vm10.local ceph-mon[51861]: pgmap v2645: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:29 vm08.local ceph-mon[47734]: pgmap v2645: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:31 vm08.local ceph-mon[47734]: pgmap v2646: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:31 vm10.local ceph-mon[51861]: pgmap v2646: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:31.693 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:31.694 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:31.722 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:31.723 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:33 vm08.local ceph-mon[47734]: pgmap v2647: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:33 vm10.local ceph-mon[51861]: pgmap v2647: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:34.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:34 vm08.local ceph-mon[47734]: pgmap v2648: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:34 vm10.local ceph-mon[51861]: pgmap v2648: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:36.724 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:36.725 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:36.751 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:36.751 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:37 vm08.local ceph-mon[47734]: pgmap v2649: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:37 vm10.local ceph-mon[51861]: pgmap v2649: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:39 vm10.local ceph-mon[51861]: pgmap v2650: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:39 vm08.local ceph-mon[47734]: pgmap v2650: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:41 vm10.local ceph-mon[51861]: pgmap v2651: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:41 vm08.local ceph-mon[47734]: pgmap v2651: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:41.752 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:41.753 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:41.781 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:41.782 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:43 vm10.local ceph-mon[51861]: pgmap v2652: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:43 vm08.local ceph-mon[47734]: pgmap v2652: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:21:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:21:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:21:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:21:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:45 vm10.local ceph-mon[51861]: pgmap v2653: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:45 vm08.local ceph-mon[47734]: pgmap v2653: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:46.783 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:46.784 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:46.814 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:46.815 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:47 vm10.local ceph-mon[51861]: pgmap v2654: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:47 vm08.local ceph-mon[47734]: pgmap v2654: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:49 vm10.local ceph-mon[51861]: pgmap v2655: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:49 vm08.local ceph-mon[47734]: pgmap v2655: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:51 vm10.local ceph-mon[51861]: pgmap v2656: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:51 vm08.local ceph-mon[47734]: pgmap v2656: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:51.816 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:51.817 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:51.843 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:51.843 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:53 vm10.local ceph-mon[51861]: pgmap v2657: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:53 vm08.local ceph-mon[47734]: pgmap v2657: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:54.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:54 vm08.local ceph-mon[47734]: pgmap v2658: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:54 vm10.local ceph-mon[51861]: pgmap v2658: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:21:56.845 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:21:56.845 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:21:56.871 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:21:56.872 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:21:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:57 vm08.local ceph-mon[47734]: pgmap v2659: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:57 vm10.local ceph-mon[51861]: pgmap v2659: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:21:59 vm08.local ceph-mon[47734]: pgmap v2660: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:21:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:21:59 vm10.local ceph-mon[51861]: pgmap v2660: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:01 vm08.local ceph-mon[47734]: pgmap v2661: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:01 vm10.local ceph-mon[51861]: pgmap v2661: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:01.873 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:01.873 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:01.902 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:01.902 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:03.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:03 vm08.local ceph-mon[47734]: pgmap v2662: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:03 vm10.local ceph-mon[51861]: pgmap v2662: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:04 vm08.local ceph-mon[47734]: pgmap v2663: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:04 vm10.local ceph-mon[51861]: pgmap v2663: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:06.904 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:06.904 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:06.931 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:06.932 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:07 vm08.local ceph-mon[47734]: pgmap v2664: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:07 vm10.local ceph-mon[51861]: pgmap v2664: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:09 vm08.local ceph-mon[47734]: pgmap v2665: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:09 vm10.local ceph-mon[51861]: pgmap v2665: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:11 vm08.local ceph-mon[47734]: pgmap v2666: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:11 vm10.local ceph-mon[51861]: pgmap v2666: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:11.933 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:11.934 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:11.960 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:11.961 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:13.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:13 vm08.local ceph-mon[47734]: pgmap v2667: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:13 vm10.local ceph-mon[51861]: pgmap v2667: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:14 vm08.local ceph-mon[47734]: pgmap v2668: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:14 vm10.local ceph-mon[51861]: pgmap v2668: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:16.962 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:16.962 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:16.995 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:16.996 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:17 vm08.local ceph-mon[47734]: pgmap v2669: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:17 vm10.local ceph-mon[51861]: pgmap v2669: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:19.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:19 vm08.local ceph-mon[47734]: pgmap v2670: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:19 vm10.local ceph-mon[51861]: pgmap v2670: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:21 vm08.local ceph-mon[47734]: pgmap v2671: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:21 vm10.local ceph-mon[51861]: pgmap v2671: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:21.997 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:21.998 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:22.024 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:22.025 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:23.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:23 vm08.local ceph-mon[47734]: pgmap v2672: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:23 vm10.local ceph-mon[51861]: pgmap v2672: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:22:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:22:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:22:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:22:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:22:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:24 vm08.local ceph-mon[47734]: pgmap v2673: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:22:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:22:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:22:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:22:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:22:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:24 vm10.local ceph-mon[51861]: pgmap v2673: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:27.027 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:27.028 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:27.055 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:27.056 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:27 vm08.local ceph-mon[47734]: pgmap v2674: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:27 vm10.local ceph-mon[51861]: pgmap v2674: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:29.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:29 vm08.local ceph-mon[47734]: pgmap v2675: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:29 vm10.local ceph-mon[51861]: pgmap v2675: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:31 vm08.local ceph-mon[47734]: pgmap v2676: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:31 vm10.local ceph-mon[51861]: pgmap v2676: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:32.057 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:32.058 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:32.084 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:32.084 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:33.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:33 vm08.local ceph-mon[47734]: pgmap v2677: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:33 vm10.local ceph-mon[51861]: pgmap v2677: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:34.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:34 vm08.local ceph-mon[47734]: pgmap v2678: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:34 vm10.local ceph-mon[51861]: pgmap v2678: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:37.086 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:37.086 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:37.112 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:37.113 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:37 vm08.local ceph-mon[47734]: pgmap v2679: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:37 vm10.local ceph-mon[51861]: pgmap v2679: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:39.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:39 vm08.local ceph-mon[47734]: pgmap v2680: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:39 vm10.local ceph-mon[51861]: pgmap v2680: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:41 vm08.local ceph-mon[47734]: pgmap v2681: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:41 vm10.local ceph-mon[51861]: pgmap v2681: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:42.114 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:42.115 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:42.142 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:42.143 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:43 vm10.local ceph-mon[51861]: pgmap v2682: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:43 vm08.local ceph-mon[47734]: pgmap v2682: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:22:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:22:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:22:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:22:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:45 vm10.local ceph-mon[51861]: pgmap v2683: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:45 vm08.local ceph-mon[47734]: pgmap v2683: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:47.145 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:47.145 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:47.171 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:47.172 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:47 vm10.local ceph-mon[51861]: pgmap v2684: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:47 vm08.local ceph-mon[47734]: pgmap v2684: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:49 vm10.local ceph-mon[51861]: pgmap v2685: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:49 vm08.local ceph-mon[47734]: pgmap v2685: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:51 vm08.local ceph-mon[47734]: pgmap v2686: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:51 vm10.local ceph-mon[51861]: pgmap v2686: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:52.174 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:52.174 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:52.200 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:52.201 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:53 vm08.local ceph-mon[47734]: pgmap v2687: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:53 vm10.local ceph-mon[51861]: pgmap v2687: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:54.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:54 vm08.local ceph-mon[47734]: pgmap v2688: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:54 vm10.local ceph-mon[51861]: pgmap v2688: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:22:57.202 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:22:57.204 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:22:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:57 vm08.local ceph-mon[47734]: pgmap v2689: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:57.368 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:22:57.368 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:22:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:57 vm10.local ceph-mon[51861]: pgmap v2689: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:59.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:22:59 vm08.local ceph-mon[47734]: pgmap v2690: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:22:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:22:59 vm10.local ceph-mon[51861]: pgmap v2690: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:01 vm10.local ceph-mon[51861]: pgmap v2691: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:01 vm08.local ceph-mon[47734]: pgmap v2691: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:02.370 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:02.371 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:02.487 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:02.487 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:03 vm10.local ceph-mon[51861]: pgmap v2692: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:03 vm08.local ceph-mon[47734]: pgmap v2692: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:04.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:04 vm08.local ceph-mon[47734]: pgmap v2693: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:04 vm10.local ceph-mon[51861]: pgmap v2693: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:07.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:07 vm08.local ceph-mon[47734]: pgmap v2694: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:07 vm10.local ceph-mon[51861]: pgmap v2694: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:07.489 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:07.489 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:07.519 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:07.520 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:09.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:09 vm08.local ceph-mon[47734]: pgmap v2695: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:09 vm10.local ceph-mon[51861]: pgmap v2695: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:11 vm10.local ceph-mon[51861]: pgmap v2696: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:11 vm08.local ceph-mon[47734]: pgmap v2696: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:12.521 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:12.522 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:12.547 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:12.548 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:13 vm10.local ceph-mon[51861]: pgmap v2697: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:13 vm08.local ceph-mon[47734]: pgmap v2697: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:14.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:14 vm08.local ceph-mon[47734]: pgmap v2698: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:14 vm10.local ceph-mon[51861]: pgmap v2698: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:17 vm10.local ceph-mon[51861]: pgmap v2699: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:17.549 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:17.550 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:17.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:17 vm08.local ceph-mon[47734]: pgmap v2699: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:17.596 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:17.597 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:19 vm10.local ceph-mon[51861]: pgmap v2700: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:19 vm08.local ceph-mon[47734]: pgmap v2700: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:21 vm08.local ceph-mon[47734]: pgmap v2701: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:21 vm10.local ceph-mon[51861]: pgmap v2701: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:22.598 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:22.599 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:22.750 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:22.774 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:23 vm08.local ceph-mon[47734]: pgmap v2702: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:23 vm10.local ceph-mon[51861]: pgmap v2702: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:23:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:23:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:23:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:24 vm08.local ceph-mon[47734]: pgmap v2703: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:23:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:23:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:23:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:23:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:23:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:24 vm10.local ceph-mon[51861]: pgmap v2703: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:23:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:23:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:27 vm08.local ceph-mon[47734]: pgmap v2704: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:27 vm10.local ceph-mon[51861]: pgmap v2704: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:27.752 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:27.752 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:27.780 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:27.780 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:29 vm10.local ceph-mon[51861]: pgmap v2705: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:29 vm08.local ceph-mon[47734]: pgmap v2705: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:31 vm08.local ceph-mon[47734]: pgmap v2706: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:31 vm10.local ceph-mon[51861]: pgmap v2706: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:32.782 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:32.782 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:32.809 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:32.809 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:33 vm08.local ceph-mon[47734]: pgmap v2707: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:33 vm10.local ceph-mon[51861]: pgmap v2707: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:34 vm08.local ceph-mon[47734]: pgmap v2708: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:34 vm10.local ceph-mon[51861]: pgmap v2708: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:37.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:37 vm08.local ceph-mon[47734]: pgmap v2709: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:37 vm10.local ceph-mon[51861]: pgmap v2709: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:37.811 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:37.811 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:37.838 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:37.838 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:39 vm08.local ceph-mon[47734]: pgmap v2710: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:39 vm10.local ceph-mon[51861]: pgmap v2710: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:41 vm08.local ceph-mon[47734]: pgmap v2711: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:41 vm10.local ceph-mon[51861]: pgmap v2711: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:42.840 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:42.841 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:42.870 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:42.870 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:43 vm08.local ceph-mon[47734]: pgmap v2712: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:23:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:23:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:43 vm10.local ceph-mon[51861]: pgmap v2712: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:23:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:23:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:44 vm08.local ceph-mon[47734]: pgmap v2713: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:44.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:44 vm10.local ceph-mon[51861]: pgmap v2713: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:47 vm08.local ceph-mon[47734]: pgmap v2714: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:47.611 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:47 vm10.local ceph-mon[51861]: pgmap v2714: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:47.872 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:47.872 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:47.925 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:47.926 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:49 vm10.local ceph-mon[51861]: pgmap v2715: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:49 vm08.local ceph-mon[47734]: pgmap v2715: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:51 vm08.local ceph-mon[47734]: pgmap v2716: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:51 vm10.local ceph-mon[51861]: pgmap v2716: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:52.927 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:52.928 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:52.956 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:52.957 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:53 vm08.local ceph-mon[47734]: pgmap v2717: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:53 vm10.local ceph-mon[51861]: pgmap v2717: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:54 vm10.local ceph-mon[51861]: pgmap v2718: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:54 vm08.local ceph-mon[47734]: pgmap v2718: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:23:57.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:57 vm08.local ceph-mon[47734]: pgmap v2719: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:57 vm10.local ceph-mon[51861]: pgmap v2719: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:57.958 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:23:57.959 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:23:57.985 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:23:57.986 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:23:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:23:59 vm10.local ceph-mon[51861]: pgmap v2720: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:23:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:23:59 vm08.local ceph-mon[47734]: pgmap v2720: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:01 vm10.local ceph-mon[51861]: pgmap v2721: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:01 vm08.local ceph-mon[47734]: pgmap v2721: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:02.987 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:02.988 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:03.017 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:03.017 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:03 vm10.local ceph-mon[51861]: pgmap v2722: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:03 vm08.local ceph-mon[47734]: pgmap v2722: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:04 vm10.local ceph-mon[51861]: pgmap v2723: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:04 vm08.local ceph-mon[47734]: pgmap v2723: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:07 vm10.local ceph-mon[51861]: pgmap v2724: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:07 vm08.local ceph-mon[47734]: pgmap v2724: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:08.019 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:08.019 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:08.046 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:08.047 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:09 vm10.local ceph-mon[51861]: pgmap v2725: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:09 vm08.local ceph-mon[47734]: pgmap v2725: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:11 vm10.local ceph-mon[51861]: pgmap v2726: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:11 vm08.local ceph-mon[47734]: pgmap v2726: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:13.048 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:13.049 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:13.107 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:13.107 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:13 vm10.local ceph-mon[51861]: pgmap v2727: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:13 vm08.local ceph-mon[47734]: pgmap v2727: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:14 vm10.local ceph-mon[51861]: pgmap v2728: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:15.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:14 vm08.local ceph-mon[47734]: pgmap v2728: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:17 vm08.local ceph-mon[47734]: pgmap v2729: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:17 vm10.local ceph-mon[51861]: pgmap v2729: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:18.109 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:18.109 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:18.137 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:18.137 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:19 vm10.local ceph-mon[51861]: pgmap v2730: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:19 vm08.local ceph-mon[47734]: pgmap v2730: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:21 vm10.local ceph-mon[51861]: pgmap v2731: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:21 vm08.local ceph-mon[47734]: pgmap v2731: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:23.139 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:23.139 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:23.166 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:23.167 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:23 vm10.local ceph-mon[51861]: pgmap v2732: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:23 vm08.local ceph-mon[47734]: pgmap v2732: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:24 vm10.local ceph-mon[51861]: pgmap v2733: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:24:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:24:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:24:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:24:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:24:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:24 vm08.local ceph-mon[47734]: pgmap v2733: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:24:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:24:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:24:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:24:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:24:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:27 vm10.local ceph-mon[51861]: pgmap v2734: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:27 vm08.local ceph-mon[47734]: pgmap v2734: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:28.169 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:28.169 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:28.197 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:28.198 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:29 vm10.local ceph-mon[51861]: pgmap v2735: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:29 vm08.local ceph-mon[47734]: pgmap v2735: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:31 vm10.local ceph-mon[51861]: pgmap v2736: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:31 vm08.local ceph-mon[47734]: pgmap v2736: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:33.200 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:33.201 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:33.233 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:33.234 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:33 vm10.local ceph-mon[51861]: pgmap v2737: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:33 vm08.local ceph-mon[47734]: pgmap v2737: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:34 vm10.local ceph-mon[51861]: pgmap v2738: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:34 vm08.local ceph-mon[47734]: pgmap v2738: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:37 vm10.local ceph-mon[51861]: pgmap v2739: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:37 vm08.local ceph-mon[47734]: pgmap v2739: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:38.235 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:38.235 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:38.261 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:38.261 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:39 vm10.local ceph-mon[51861]: pgmap v2740: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:39 vm08.local ceph-mon[47734]: pgmap v2740: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:41 vm10.local ceph-mon[51861]: pgmap v2741: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:41 vm08.local ceph-mon[47734]: pgmap v2741: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:43.262 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:43.263 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:43.290 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:43.290 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:43 vm10.local ceph-mon[51861]: pgmap v2742: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:43 vm08.local ceph-mon[47734]: pgmap v2742: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:24:44.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:24:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:24:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:24:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:45 vm10.local ceph-mon[51861]: pgmap v2743: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:45 vm08.local ceph-mon[47734]: pgmap v2743: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:47 vm10.local ceph-mon[51861]: pgmap v2744: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:47 vm08.local ceph-mon[47734]: pgmap v2744: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:48.292 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:48.297 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:48.473 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:48.474 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:49 vm08.local ceph-mon[47734]: pgmap v2745: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:49 vm10.local ceph-mon[51861]: pgmap v2745: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:51 vm08.local ceph-mon[47734]: pgmap v2746: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:51 vm10.local ceph-mon[51861]: pgmap v2746: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:53.475 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:53.476 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:53.504 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:53.505 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:53 vm08.local ceph-mon[47734]: pgmap v2747: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:53 vm10.local ceph-mon[51861]: pgmap v2747: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:54 vm10.local ceph-mon[51861]: pgmap v2748: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:54 vm08.local ceph-mon[47734]: pgmap v2748: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:24:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:57 vm10.local ceph-mon[51861]: pgmap v2749: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:57 vm08.local ceph-mon[47734]: pgmap v2749: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:58.507 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:24:58.507 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:24:58.534 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:24:58.535 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:24:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:24:59 vm10.local ceph-mon[51861]: pgmap v2750: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:24:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:24:59 vm08.local ceph-mon[47734]: pgmap v2750: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:01 vm10.local ceph-mon[51861]: pgmap v2751: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:01 vm08.local ceph-mon[47734]: pgmap v2751: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:03 vm10.local ceph-mon[51861]: pgmap v2752: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:03.536 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:03.536 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:03 vm08.local ceph-mon[47734]: pgmap v2752: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:03.561 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:03.562 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:04 vm10.local ceph-mon[51861]: pgmap v2753: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:04 vm08.local ceph-mon[47734]: pgmap v2753: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:07 vm10.local ceph-mon[51861]: pgmap v2754: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:07 vm08.local ceph-mon[47734]: pgmap v2754: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:08.563 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:08.564 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:08.590 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:08.591 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:09 vm10.local ceph-mon[51861]: pgmap v2755: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:09 vm08.local ceph-mon[47734]: pgmap v2755: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:11 vm10.local ceph-mon[51861]: pgmap v2756: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:11 vm08.local ceph-mon[47734]: pgmap v2756: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:13 vm10.local ceph-mon[51861]: pgmap v2757: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:13 vm08.local ceph-mon[47734]: pgmap v2757: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:13.592 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:13.593 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:13.619 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:13.620 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:14 vm10.local ceph-mon[51861]: pgmap v2758: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:14 vm08.local ceph-mon[47734]: pgmap v2758: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:17 vm10.local ceph-mon[51861]: pgmap v2759: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:17 vm08.local ceph-mon[47734]: pgmap v2759: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:18.621 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:18.622 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:18.649 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:18.649 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:19 vm10.local ceph-mon[51861]: pgmap v2760: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:19 vm08.local ceph-mon[47734]: pgmap v2760: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:21 vm10.local ceph-mon[51861]: pgmap v2761: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:21 vm08.local ceph-mon[47734]: pgmap v2761: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:23 vm10.local ceph-mon[51861]: pgmap v2762: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:23 vm08.local ceph-mon[47734]: pgmap v2762: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:23.651 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:23.651 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:23.710 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:23.710 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:24.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:24 vm08.local ceph-mon[47734]: pgmap v2763: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:25:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:25:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:25:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:24.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:24 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:24 vm10.local ceph-mon[51861]: pgmap v2763: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:25:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:25:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:25:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:24 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:26.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:25 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:26.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:25 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:26.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:25 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:26.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:25 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:26.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:25 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:26.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:25 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:26.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:25 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:26.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:25 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:25:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:26 vm10.local ceph-mon[51861]: pgmap v2764: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:27.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:26 vm08.local ceph-mon[47734]: pgmap v2764: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:28.711 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:28.712 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:28.739 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:28.739 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:29 vm10.local ceph-mon[51861]: pgmap v2765: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:29 vm08.local ceph-mon[47734]: pgmap v2765: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:31 vm10.local ceph-mon[51861]: pgmap v2766: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:31 vm08.local ceph-mon[47734]: pgmap v2766: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:33 vm10.local ceph-mon[51861]: pgmap v2767: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:25:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:33 vm08.local ceph-mon[47734]: pgmap v2767: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:25:33.740 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:33.741 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:33.768 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:33.769 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:34 vm10.local ceph-mon[51861]: pgmap v2768: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:34 vm08.local ceph-mon[47734]: pgmap v2768: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:37 vm10.local ceph-mon[51861]: pgmap v2769: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:25:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:37 vm08.local ceph-mon[47734]: pgmap v2769: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:25:38.770 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:38.771 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:38.797 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:38.797 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:39 vm10.local ceph-mon[51861]: pgmap v2770: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:25:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:39 vm08.local ceph-mon[47734]: pgmap v2770: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:25:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:41 vm10.local ceph-mon[51861]: pgmap v2771: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:41 vm08.local ceph-mon[47734]: pgmap v2771: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:43 vm10.local ceph-mon[51861]: pgmap v2772: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:43 vm08.local ceph-mon[47734]: pgmap v2772: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:43.798 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:43.799 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:43.826 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:43.827 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:25:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:25:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:25:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:25:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:45 vm10.local ceph-mon[51861]: pgmap v2773: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:45 vm08.local ceph-mon[47734]: pgmap v2773: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:47.351 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:47 vm10.local ceph-mon[51861]: pgmap v2774: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:47 vm08.local ceph-mon[47734]: pgmap v2774: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:48.828 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:48.828 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:48.855 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:48.856 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:49 vm10.local ceph-mon[51861]: pgmap v2775: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:49 vm08.local ceph-mon[47734]: pgmap v2775: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:51 vm10.local ceph-mon[51861]: pgmap v2776: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:51 vm08.local ceph-mon[47734]: pgmap v2776: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:53 vm10.local ceph-mon[51861]: pgmap v2777: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:53 vm08.local ceph-mon[47734]: pgmap v2777: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:53.857 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:53.857 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:53.885 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:53.886 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:54.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:54 vm10.local ceph-mon[51861]: pgmap v2778: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:54 vm08.local ceph-mon[47734]: pgmap v2778: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:25:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:57 vm10.local ceph-mon[51861]: pgmap v2779: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:57 vm08.local ceph-mon[47734]: pgmap v2779: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:58.887 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:25:58.887 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:25:58.913 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:25:58.914 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:25:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:25:59 vm10.local ceph-mon[51861]: pgmap v2780: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:25:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:25:59 vm08.local ceph-mon[47734]: pgmap v2780: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:01 vm10.local ceph-mon[51861]: pgmap v2781: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:01 vm08.local ceph-mon[47734]: pgmap v2781: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:03 vm10.local ceph-mon[51861]: pgmap v2782: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:03 vm08.local ceph-mon[47734]: pgmap v2782: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:03.916 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:03.916 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:03.944 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:03.945 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:04.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:04 vm10.local ceph-mon[51861]: pgmap v2783: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:04 vm08.local ceph-mon[47734]: pgmap v2783: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:07 vm10.local ceph-mon[51861]: pgmap v2784: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:07.495 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:07 vm08.local ceph-mon[47734]: pgmap v2784: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:08.947 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:08.947 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:08.975 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:08.975 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:09 vm10.local ceph-mon[51861]: pgmap v2785: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:09 vm08.local ceph-mon[47734]: pgmap v2785: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:11 vm10.local ceph-mon[51861]: pgmap v2786: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:11 vm08.local ceph-mon[47734]: pgmap v2786: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:13 vm10.local ceph-mon[51861]: pgmap v2787: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:13 vm08.local ceph-mon[47734]: pgmap v2787: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:13.977 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:13.977 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:14.003 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:14.004 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:14.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:14 vm10.local ceph-mon[51861]: pgmap v2788: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:14 vm08.local ceph-mon[47734]: pgmap v2788: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:17 vm08.local ceph-mon[47734]: pgmap v2789: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:17.611 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:17 vm10.local ceph-mon[51861]: pgmap v2789: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:19.006 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:19.006 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:19.165 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:19.166 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:19 vm08.local ceph-mon[47734]: pgmap v2790: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:19 vm10.local ceph-mon[51861]: pgmap v2790: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:21 vm08.local ceph-mon[47734]: pgmap v2791: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:21 vm10.local ceph-mon[51861]: pgmap v2791: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:23 vm08.local ceph-mon[47734]: pgmap v2792: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:23 vm10.local ceph-mon[51861]: pgmap v2792: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:24.167 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:24.168 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:24.193 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:24.193 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:24.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:24 vm10.local ceph-mon[51861]: pgmap v2793: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:24.983 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:24 vm08.local ceph-mon[47734]: pgmap v2793: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:25.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:25 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:26:25.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:25 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:26:25.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:25 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:26:25.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:25 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:26:25.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:25 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:26:25.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:25 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:26:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:26 vm08.local ceph-mon[47734]: pgmap v2794: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:26 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:26:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:26 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:26:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:26 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:26:26.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:26 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:26:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:26 vm10.local ceph-mon[51861]: pgmap v2794: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:26 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:26:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:26 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:26:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:26 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:26:27.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:26 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:26:29.195 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:29.196 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:29.223 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:29.224 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:29 vm10.local ceph-mon[51861]: pgmap v2795: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:29 vm08.local ceph-mon[47734]: pgmap v2795: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:31 vm10.local ceph-mon[51861]: pgmap v2796: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:31 vm08.local ceph-mon[47734]: pgmap v2796: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:33 vm10.local ceph-mon[51861]: pgmap v2797: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:33 vm08.local ceph-mon[47734]: pgmap v2797: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:34.225 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:34.226 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:34.252 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:34.253 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:35.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:34 vm08.local ceph-mon[47734]: pgmap v2798: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:34 vm10.local ceph-mon[51861]: pgmap v2798: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:37 vm10.local ceph-mon[51861]: pgmap v2799: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:37 vm08.local ceph-mon[47734]: pgmap v2799: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:39.255 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:39.255 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:39.280 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:39.280 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:39 vm10.local ceph-mon[51861]: pgmap v2800: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:39 vm08.local ceph-mon[47734]: pgmap v2800: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:41 vm10.local ceph-mon[51861]: pgmap v2801: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:41 vm08.local ceph-mon[47734]: pgmap v2801: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:43 vm10.local ceph-mon[51861]: pgmap v2802: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:43 vm08.local ceph-mon[47734]: pgmap v2802: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:44.282 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:44.282 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:26:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:26:44.310 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:44.311 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:26:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:26:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:45 vm10.local ceph-mon[51861]: pgmap v2803: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:45 vm08.local ceph-mon[47734]: pgmap v2803: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:47 vm10.local ceph-mon[51861]: pgmap v2804: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:47 vm08.local ceph-mon[47734]: pgmap v2804: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:49.312 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:49.312 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:49.339 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:49.340 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:49 vm10.local ceph-mon[51861]: pgmap v2805: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:49 vm08.local ceph-mon[47734]: pgmap v2805: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:51 vm10.local ceph-mon[51861]: pgmap v2806: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:51 vm08.local ceph-mon[47734]: pgmap v2806: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:53 vm08.local ceph-mon[47734]: pgmap v2807: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:53 vm10.local ceph-mon[51861]: pgmap v2807: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:54.341 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:54.341 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:54.369 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:54.370 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:55.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:54 vm08.local ceph-mon[47734]: pgmap v2808: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:54 vm10.local ceph-mon[51861]: pgmap v2808: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:26:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:57 vm10.local ceph-mon[51861]: pgmap v2809: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:57 vm08.local ceph-mon[47734]: pgmap v2809: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:59.371 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:26:59.372 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:26:59.398 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:26:59.399 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:26:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:26:59 vm10.local ceph-mon[51861]: pgmap v2810: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:26:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:26:59 vm08.local ceph-mon[47734]: pgmap v2810: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:01 vm10.local ceph-mon[51861]: pgmap v2811: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:01 vm08.local ceph-mon[47734]: pgmap v2811: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:03 vm10.local ceph-mon[51861]: pgmap v2812: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:03 vm08.local ceph-mon[47734]: pgmap v2812: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:04.400 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:04.401 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:04.427 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:04.427 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:05.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:04 vm08.local ceph-mon[47734]: pgmap v2813: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:04 vm10.local ceph-mon[51861]: pgmap v2813: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:07 vm10.local ceph-mon[51861]: pgmap v2814: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:07 vm08.local ceph-mon[47734]: pgmap v2814: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:09 vm10.local ceph-mon[51861]: pgmap v2815: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:09.429 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:09.429 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:09.455 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:09.456 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:09 vm08.local ceph-mon[47734]: pgmap v2815: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:11 vm10.local ceph-mon[51861]: pgmap v2816: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:11 vm08.local ceph-mon[47734]: pgmap v2816: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:13 vm10.local ceph-mon[51861]: pgmap v2817: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:13 vm08.local ceph-mon[47734]: pgmap v2817: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:14.457 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:14.458 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:14.485 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:14.486 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:15.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:14 vm08.local ceph-mon[47734]: pgmap v2818: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:14 vm10.local ceph-mon[51861]: pgmap v2818: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:17 vm10.local ceph-mon[51861]: pgmap v2819: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:17 vm08.local ceph-mon[47734]: pgmap v2819: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:19 vm10.local ceph-mon[51861]: pgmap v2820: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:19.487 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:19.488 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:19.514 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:19.514 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:19 vm08.local ceph-mon[47734]: pgmap v2820: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:21 vm08.local ceph-mon[47734]: pgmap v2821: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:21 vm10.local ceph-mon[51861]: pgmap v2821: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:23 vm08.local ceph-mon[47734]: pgmap v2822: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:23 vm10.local ceph-mon[51861]: pgmap v2822: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:24.516 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:24.516 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:24.546 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:24.546 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:24 vm08.local ceph-mon[47734]: pgmap v2823: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:24 vm10.local ceph-mon[51861]: pgmap v2823: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:27.257 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:27 vm08.local ceph-mon[47734]: pgmap v2824: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:27.257 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:27 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:27:27.257 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:27 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:27:27.257 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:27 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:27:27.375 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:27 vm10.local ceph-mon[51861]: pgmap v2824: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:27.375 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:27 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:27:27.375 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:27 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:27:27.375 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:27 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:27:28.452 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:28 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:28.452 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:28 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:28.452 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:28 vm08.local ceph-mon[47734]: pgmap v2825: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:28.848 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:28 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:28.848 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:28 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:28.848 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:28 vm10.local ceph-mon[51861]: pgmap v2825: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:29.547 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:29.548 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:29.575 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:29.576 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:27:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:27:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:27:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:30 vm08.local ceph-mon[47734]: pgmap v2826: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:30 vm10.local ceph-mon[51861]: pgmap v2826: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:33 vm10.local ceph-mon[51861]: pgmap v2827: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:33 vm08.local ceph-mon[47734]: pgmap v2827: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:34.577 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:34.577 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:34.604 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:34.605 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:34 vm10.local ceph-mon[51861]: pgmap v2828: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:34 vm08.local ceph-mon[47734]: pgmap v2828: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:37 vm08.local ceph-mon[47734]: pgmap v2829: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:37 vm10.local ceph-mon[51861]: pgmap v2829: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:39 vm10.local ceph-mon[51861]: pgmap v2830: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:39 vm08.local ceph-mon[47734]: pgmap v2830: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:39.606 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:39.606 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:39.632 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:39.633 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:41 vm08.local ceph-mon[47734]: pgmap v2831: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:41 vm10.local ceph-mon[51861]: pgmap v2831: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:43 vm08.local ceph-mon[47734]: pgmap v2832: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:27:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:27:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:43 vm10.local ceph-mon[51861]: pgmap v2832: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:27:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:27:44.634 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:44.635 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:44.662 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:44.662 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:44 vm10.local ceph-mon[51861]: pgmap v2833: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:44 vm08.local ceph-mon[47734]: pgmap v2833: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:47 vm10.local ceph-mon[51861]: pgmap v2834: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:47 vm08.local ceph-mon[47734]: pgmap v2834: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:49 vm10.local ceph-mon[51861]: pgmap v2835: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:49 vm08.local ceph-mon[47734]: pgmap v2835: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:49.663 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:49.664 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:49.690 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:49.690 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:51 vm10.local ceph-mon[51861]: pgmap v2836: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:51 vm08.local ceph-mon[47734]: pgmap v2836: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:53 vm10.local ceph-mon[51861]: pgmap v2837: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:53 vm08.local ceph-mon[47734]: pgmap v2837: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:54.692 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:54.692 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:54.718 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:54.719 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:27:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:54 vm10.local ceph-mon[51861]: pgmap v2838: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:54 vm08.local ceph-mon[47734]: pgmap v2838: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:27:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:57 vm10.local ceph-mon[51861]: pgmap v2839: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:57 vm08.local ceph-mon[47734]: pgmap v2839: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:27:59 vm10.local ceph-mon[51861]: pgmap v2840: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:27:59 vm08.local ceph-mon[47734]: pgmap v2840: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:27:59.720 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:27:59.720 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:27:59.748 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:27:59.748 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:01 vm10.local ceph-mon[51861]: pgmap v2841: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:01 vm08.local ceph-mon[47734]: pgmap v2841: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:03 vm10.local ceph-mon[51861]: pgmap v2842: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:03 vm08.local ceph-mon[47734]: pgmap v2842: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:04.750 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:04.750 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:04.776 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:04.777 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:04 vm10.local ceph-mon[51861]: pgmap v2843: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:04 vm08.local ceph-mon[47734]: pgmap v2843: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:07 vm10.local ceph-mon[51861]: pgmap v2844: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:07.508 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:07 vm08.local ceph-mon[47734]: pgmap v2844: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:09 vm10.local ceph-mon[51861]: pgmap v2845: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:09 vm08.local ceph-mon[47734]: pgmap v2845: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:09.778 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:09.779 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:09.807 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:09.807 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:11 vm10.local ceph-mon[51861]: pgmap v2846: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:11 vm08.local ceph-mon[47734]: pgmap v2846: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:13 vm10.local ceph-mon[51861]: pgmap v2847: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:13 vm08.local ceph-mon[47734]: pgmap v2847: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:14.809 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:14.810 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:14.845 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:14.846 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:14 vm10.local ceph-mon[51861]: pgmap v2848: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:14 vm08.local ceph-mon[47734]: pgmap v2848: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:17 vm10.local ceph-mon[51861]: pgmap v2849: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:17.554 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:17 vm08.local ceph-mon[47734]: pgmap v2849: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:19 vm10.local ceph-mon[51861]: pgmap v2850: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:19 vm08.local ceph-mon[47734]: pgmap v2850: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:19.847 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:19.848 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:19.875 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:19.876 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:21 vm10.local ceph-mon[51861]: pgmap v2851: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:21 vm08.local ceph-mon[47734]: pgmap v2851: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:23 vm10.local ceph-mon[51861]: pgmap v2852: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:23 vm08.local ceph-mon[47734]: pgmap v2852: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:24.877 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:24.878 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:24.905 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:24.906 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:24 vm10.local ceph-mon[51861]: pgmap v2853: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:24 vm08.local ceph-mon[47734]: pgmap v2853: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:27 vm10.local ceph-mon[51861]: pgmap v2854: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:27 vm08.local ceph-mon[47734]: pgmap v2854: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:29.192 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:29 vm10.local ceph-mon[51861]: pgmap v2855: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:29.192 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:28:29.192 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:28:29.192 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:28:29.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:29 vm08.local ceph-mon[47734]: pgmap v2855: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:29.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:28:29.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:28:29.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:28:29.907 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:29.907 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:29.938 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:29.939 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:28:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:28:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:28:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:28:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:31 vm08.local ceph-mon[47734]: pgmap v2856: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:31 vm10.local ceph-mon[51861]: pgmap v2856: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:33 vm08.local ceph-mon[47734]: pgmap v2857: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:33 vm10.local ceph-mon[51861]: pgmap v2857: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:34.940 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:34.941 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:34.967 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:34.968 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:34 vm10.local ceph-mon[51861]: pgmap v2858: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:34 vm08.local ceph-mon[47734]: pgmap v2858: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:37 vm10.local ceph-mon[51861]: pgmap v2859: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:37 vm08.local ceph-mon[47734]: pgmap v2859: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:39 vm10.local ceph-mon[51861]: pgmap v2860: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:39 vm08.local ceph-mon[47734]: pgmap v2860: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:39.969 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:39.970 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:39.995 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:39.996 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:41 vm10.local ceph-mon[51861]: pgmap v2861: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:41 vm08.local ceph-mon[47734]: pgmap v2861: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:43 vm10.local ceph-mon[51861]: pgmap v2862: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:28:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:43 vm08.local ceph-mon[47734]: pgmap v2862: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:28:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:28:44.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:28:44.998 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:44.998 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:45.060 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:45.061 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:45 vm10.local ceph-mon[51861]: pgmap v2863: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:45 vm08.local ceph-mon[47734]: pgmap v2863: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:47 vm10.local ceph-mon[51861]: pgmap v2864: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:47 vm08.local ceph-mon[47734]: pgmap v2864: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:49 vm10.local ceph-mon[51861]: pgmap v2865: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:49 vm08.local ceph-mon[47734]: pgmap v2865: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:50.062 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:50.062 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:50.090 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:50.090 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:51 vm10.local ceph-mon[51861]: pgmap v2866: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:51 vm08.local ceph-mon[47734]: pgmap v2866: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:53 vm10.local ceph-mon[51861]: pgmap v2867: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:53 vm08.local ceph-mon[47734]: pgmap v2867: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:55.092 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:28:55.092 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:28:55.120 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:28:55.121 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:28:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:54 vm10.local ceph-mon[51861]: pgmap v2868: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:54 vm08.local ceph-mon[47734]: pgmap v2868: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:28:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:57 vm10.local ceph-mon[51861]: pgmap v2869: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:57 vm08.local ceph-mon[47734]: pgmap v2869: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:28:59 vm10.local ceph-mon[51861]: pgmap v2870: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:28:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:28:59 vm08.local ceph-mon[47734]: pgmap v2870: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:00.122 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:00.123 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:00.149 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:00.150 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:01 vm10.local ceph-mon[51861]: pgmap v2871: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:01 vm08.local ceph-mon[47734]: pgmap v2871: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:03 vm10.local ceph-mon[51861]: pgmap v2872: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:03 vm08.local ceph-mon[47734]: pgmap v2872: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:05.151 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:05.152 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:04 vm10.local ceph-mon[51861]: pgmap v2873: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:05.179 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:05.180 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:04 vm08.local ceph-mon[47734]: pgmap v2873: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:07 vm10.local ceph-mon[51861]: pgmap v2874: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:07 vm08.local ceph-mon[47734]: pgmap v2874: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:09 vm10.local ceph-mon[51861]: pgmap v2875: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:09 vm08.local ceph-mon[47734]: pgmap v2875: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:10.181 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:10.182 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:10.232 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:10.233 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:11 vm10.local ceph-mon[51861]: pgmap v2876: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:11 vm08.local ceph-mon[47734]: pgmap v2876: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:13 vm10.local ceph-mon[51861]: pgmap v2877: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:13 vm08.local ceph-mon[47734]: pgmap v2877: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:14 vm10.local ceph-mon[51861]: pgmap v2878: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:15.234 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:15.235 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:15.260 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:15.261 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:14 vm08.local ceph-mon[47734]: pgmap v2878: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:17 vm10.local ceph-mon[51861]: pgmap v2879: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:17 vm08.local ceph-mon[47734]: pgmap v2879: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:19 vm10.local ceph-mon[51861]: pgmap v2880: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:19 vm08.local ceph-mon[47734]: pgmap v2880: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:20.262 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:20.262 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:20.288 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:20.288 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:21 vm10.local ceph-mon[51861]: pgmap v2881: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:21 vm08.local ceph-mon[47734]: pgmap v2881: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:23 vm10.local ceph-mon[51861]: pgmap v2882: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:23 vm08.local ceph-mon[47734]: pgmap v2882: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:24 vm10.local ceph-mon[51861]: pgmap v2883: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:25.289 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:25.290 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:24 vm08.local ceph-mon[47734]: pgmap v2883: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:25.316 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:25.316 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:27 vm08.local ceph-mon[47734]: pgmap v2884: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:27 vm10.local ceph-mon[51861]: pgmap v2884: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:29.227 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:29 vm10.local ceph-mon[51861]: pgmap v2885: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:29.227 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:29:29.227 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:29:29.227 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:29 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:29:29.449 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:29 vm08.local ceph-mon[47734]: pgmap v2885: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:29.449 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:29:29.449 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:29:29.449 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:29 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:29:30.318 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:30.318 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:30.346 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:30.347 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:29:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:29:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:30 vm08.local ceph-mon[47734]: pgmap v2886: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:29:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:29:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:30 vm10.local ceph-mon[51861]: pgmap v2886: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:33 vm10.local ceph-mon[51861]: pgmap v2887: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:33 vm08.local ceph-mon[47734]: pgmap v2887: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:35 vm08.local ceph-mon[47734]: pgmap v2888: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:35.348 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:35.349 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:35.376 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:35.377 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:35 vm10.local ceph-mon[51861]: pgmap v2888: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:37 vm10.local ceph-mon[51861]: pgmap v2889: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:37 vm08.local ceph-mon[47734]: pgmap v2889: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:39 vm10.local ceph-mon[51861]: pgmap v2890: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:39 vm08.local ceph-mon[47734]: pgmap v2890: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:40.378 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:40.378 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:40.404 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:40.405 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:41 vm10.local ceph-mon[51861]: pgmap v2891: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:41 vm08.local ceph-mon[47734]: pgmap v2891: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:43 vm10.local ceph-mon[51861]: pgmap v2892: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:29:43.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:29:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:43 vm08.local ceph-mon[47734]: pgmap v2892: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:29:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:29:45.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:44 vm10.local ceph-mon[51861]: pgmap v2893: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:44 vm08.local ceph-mon[47734]: pgmap v2893: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:45.406 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:45.407 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:45.438 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:45.439 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:47 vm10.local ceph-mon[51861]: pgmap v2894: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:47 vm08.local ceph-mon[47734]: pgmap v2894: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:49 vm10.local ceph-mon[51861]: pgmap v2895: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:49 vm08.local ceph-mon[47734]: pgmap v2895: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:50.441 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:50.441 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:50.467 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:50.468 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:51 vm10.local ceph-mon[51861]: pgmap v2896: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:51 vm08.local ceph-mon[47734]: pgmap v2896: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:53.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:53 vm10.local ceph-mon[51861]: pgmap v2897: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:53 vm08.local ceph-mon[47734]: pgmap v2897: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:55.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:54 vm10.local ceph-mon[51861]: pgmap v2898: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:54 vm08.local ceph-mon[47734]: pgmap v2898: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:29:55.470 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:29:55.470 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:29:55.499 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:29:55.499 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:29:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:57 vm10.local ceph-mon[51861]: pgmap v2899: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:57 vm08.local ceph-mon[47734]: pgmap v2899: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:29:59 vm10.local ceph-mon[51861]: pgmap v2900: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:29:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:29:59 vm08.local ceph-mon[47734]: pgmap v2900: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:00.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T15:30:00.500 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:00.501 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:00.527 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:00.528 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:00.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T15:30:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:01 vm10.local ceph-mon[51861]: pgmap v2901: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:01 vm08.local ceph-mon[47734]: pgmap v2901: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:03.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:03 vm10.local ceph-mon[51861]: pgmap v2902: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:03 vm08.local ceph-mon[47734]: pgmap v2902: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:05.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:04 vm10.local ceph-mon[51861]: pgmap v2903: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:04 vm08.local ceph-mon[47734]: pgmap v2903: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:05.529 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:05.530 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:05.556 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:05.557 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:07 vm10.local ceph-mon[51861]: pgmap v2904: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:07 vm08.local ceph-mon[47734]: pgmap v2904: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:09 vm10.local ceph-mon[51861]: pgmap v2905: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:09 vm08.local ceph-mon[47734]: pgmap v2905: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:10.558 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:10.559 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:10.587 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:10.588 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:11 vm10.local ceph-mon[51861]: pgmap v2906: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:11 vm08.local ceph-mon[47734]: pgmap v2906: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:13.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:13 vm10.local ceph-mon[51861]: pgmap v2907: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:13 vm08.local ceph-mon[47734]: pgmap v2907: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:15.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:14 vm10.local ceph-mon[51861]: pgmap v2908: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:14 vm08.local ceph-mon[47734]: pgmap v2908: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:15.589 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:15.590 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:15.615 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:15.615 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:17 vm10.local ceph-mon[51861]: pgmap v2909: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:17 vm08.local ceph-mon[47734]: pgmap v2909: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:19 vm10.local ceph-mon[51861]: pgmap v2910: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:19 vm08.local ceph-mon[47734]: pgmap v2910: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:20.617 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:20.617 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:20.642 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:20.643 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:21 vm10.local ceph-mon[51861]: pgmap v2911: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:21 vm08.local ceph-mon[47734]: pgmap v2911: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:23.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:23 vm10.local ceph-mon[51861]: pgmap v2912: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:23 vm08.local ceph-mon[47734]: pgmap v2912: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:24 vm10.local ceph-mon[51861]: pgmap v2913: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:24 vm08.local ceph-mon[47734]: pgmap v2913: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:25.644 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:25.644 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:25.672 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:25.672 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:27 vm10.local ceph-mon[51861]: pgmap v2914: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:27 vm08.local ceph-mon[47734]: pgmap v2914: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:29 vm10.local ceph-mon[51861]: pgmap v2915: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:29.492 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:29 vm08.local ceph-mon[47734]: pgmap v2915: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:30:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:30:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:30:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:30:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:30:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:30:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:30:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:30:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:30:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:30:30.674 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:30.674 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:30.701 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:30.701 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:31 vm10.local ceph-mon[51861]: pgmap v2916: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:31 vm08.local ceph-mon[47734]: pgmap v2916: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:33.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:33 vm10.local ceph-mon[51861]: pgmap v2917: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:33 vm08.local ceph-mon[47734]: pgmap v2917: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:35.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:34 vm10.local ceph-mon[51861]: pgmap v2918: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:34 vm08.local ceph-mon[47734]: pgmap v2918: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:35.702 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:35.703 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:35.729 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:35.729 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:37 vm10.local ceph-mon[51861]: pgmap v2919: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:37 vm08.local ceph-mon[47734]: pgmap v2919: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:39 vm08.local ceph-mon[47734]: pgmap v2920: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:39 vm10.local ceph-mon[51861]: pgmap v2920: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:40.731 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:40.731 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:40.757 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:40.758 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:41 vm08.local ceph-mon[47734]: pgmap v2921: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:41 vm10.local ceph-mon[51861]: pgmap v2921: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:43 vm08.local ceph-mon[47734]: pgmap v2922: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:30:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:30:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:43 vm10.local ceph-mon[51861]: pgmap v2922: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:30:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:30:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:44 vm08.local ceph-mon[47734]: pgmap v2923: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:44 vm10.local ceph-mon[51861]: pgmap v2923: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:45.759 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:45.760 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:45.788 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:45.788 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:47 vm08.local ceph-mon[47734]: pgmap v2924: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:47 vm10.local ceph-mon[51861]: pgmap v2924: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:49 vm08.local ceph-mon[47734]: pgmap v2925: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:49 vm10.local ceph-mon[51861]: pgmap v2925: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:50.790 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:50.790 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:50.818 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:50.819 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:51 vm08.local ceph-mon[47734]: pgmap v2926: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:51 vm10.local ceph-mon[51861]: pgmap v2926: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:53 vm08.local ceph-mon[47734]: pgmap v2927: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:53 vm10.local ceph-mon[51861]: pgmap v2927: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:54 vm08.local ceph-mon[47734]: pgmap v2928: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:54 vm10.local ceph-mon[51861]: pgmap v2928: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:30:55.821 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:30:55.821 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:30:55.847 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:30:55.848 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:30:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:57 vm08.local ceph-mon[47734]: pgmap v2929: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:57 vm10.local ceph-mon[51861]: pgmap v2929: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:30:59 vm08.local ceph-mon[47734]: pgmap v2930: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:30:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:30:59 vm10.local ceph-mon[51861]: pgmap v2930: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:00.849 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:00.850 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:00.876 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:00.877 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:01 vm08.local ceph-mon[47734]: pgmap v2931: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:01 vm10.local ceph-mon[51861]: pgmap v2931: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:03 vm08.local ceph-mon[47734]: pgmap v2932: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:03 vm10.local ceph-mon[51861]: pgmap v2932: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:04 vm08.local ceph-mon[47734]: pgmap v2933: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:04 vm10.local ceph-mon[51861]: pgmap v2933: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:05.878 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:05.879 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:05.905 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:05.906 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:07 vm10.local ceph-mon[51861]: pgmap v2934: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:07 vm08.local ceph-mon[47734]: pgmap v2934: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:09 vm10.local ceph-mon[51861]: pgmap v2935: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:09 vm08.local ceph-mon[47734]: pgmap v2935: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:10.907 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:10.908 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:10.934 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:10.934 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:11 vm10.local ceph-mon[51861]: pgmap v2936: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:11 vm08.local ceph-mon[47734]: pgmap v2936: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:13 vm08.local ceph-mon[47734]: pgmap v2937: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:13 vm10.local ceph-mon[51861]: pgmap v2937: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:15.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:14 vm08.local ceph-mon[47734]: pgmap v2938: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:14 vm10.local ceph-mon[51861]: pgmap v2938: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:15.935 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:15.936 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:15.961 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:15.962 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:17 vm10.local ceph-mon[51861]: pgmap v2939: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:17 vm08.local ceph-mon[47734]: pgmap v2939: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:19 vm08.local ceph-mon[47734]: pgmap v2940: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:19 vm10.local ceph-mon[51861]: pgmap v2940: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:20.964 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:20.964 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:20.990 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:20.991 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:21 vm08.local ceph-mon[47734]: pgmap v2941: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:21 vm10.local ceph-mon[51861]: pgmap v2941: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:23 vm08.local ceph-mon[47734]: pgmap v2942: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:23 vm10.local ceph-mon[51861]: pgmap v2942: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:24 vm08.local ceph-mon[47734]: pgmap v2943: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:24 vm10.local ceph-mon[51861]: pgmap v2943: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:25.992 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:25.992 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:26.020 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:26.021 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:27.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:27 vm10.local ceph-mon[51861]: pgmap v2944: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:27 vm08.local ceph-mon[47734]: pgmap v2944: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:29 vm10.local ceph-mon[51861]: pgmap v2945: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:29 vm08.local ceph-mon[47734]: pgmap v2945: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:30.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:31:30.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:31:30.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:31:30.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:31:30.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:31:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:31:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:31:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:31:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:31:30.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:31:31.022 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:31.022 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:31.048 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:31.048 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:31 vm08.local ceph-mon[47734]: pgmap v2946: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:31 vm10.local ceph-mon[51861]: pgmap v2946: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:33 vm08.local ceph-mon[47734]: pgmap v2947: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:33 vm10.local ceph-mon[51861]: pgmap v2947: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:35.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:34 vm08.local ceph-mon[47734]: pgmap v2948: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:34 vm10.local ceph-mon[51861]: pgmap v2948: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:36.049 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:36.050 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:36.075 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:36.075 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:37 vm10.local ceph-mon[51861]: pgmap v2949: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:37 vm08.local ceph-mon[47734]: pgmap v2949: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:39 vm10.local ceph-mon[51861]: pgmap v2950: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:39 vm08.local ceph-mon[47734]: pgmap v2950: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:41.076 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:41.077 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:41.101 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:41.102 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:41 vm10.local ceph-mon[51861]: pgmap v2951: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:41 vm08.local ceph-mon[47734]: pgmap v2951: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:43 vm08.local ceph-mon[47734]: pgmap v2952: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:31:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:31:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:43 vm10.local ceph-mon[51861]: pgmap v2952: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:31:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:31:45.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:44 vm08.local ceph-mon[47734]: pgmap v2953: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:44 vm10.local ceph-mon[51861]: pgmap v2953: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:46.103 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:46.103 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:46.130 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:46.131 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:47.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:47 vm10.local ceph-mon[51861]: pgmap v2954: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:47 vm08.local ceph-mon[47734]: pgmap v2954: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:49.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:49 vm10.local ceph-mon[51861]: pgmap v2955: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:49 vm08.local ceph-mon[47734]: pgmap v2955: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:51.132 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:51.133 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:51.160 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:51.160 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:51 vm08.local ceph-mon[47734]: pgmap v2956: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:51 vm10.local ceph-mon[51861]: pgmap v2956: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:53 vm08.local ceph-mon[47734]: pgmap v2957: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:53 vm10.local ceph-mon[51861]: pgmap v2957: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:55.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:54 vm08.local ceph-mon[47734]: pgmap v2958: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:55.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:54 vm10.local ceph-mon[51861]: pgmap v2958: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:31:56.162 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:31:56.162 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:31:56.214 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:31:56.214 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:31:57.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:57 vm10.local ceph-mon[51861]: pgmap v2959: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:57 vm08.local ceph-mon[47734]: pgmap v2959: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:59.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:31:59 vm10.local ceph-mon[51861]: pgmap v2960: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:31:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:31:59 vm08.local ceph-mon[47734]: pgmap v2960: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:01.216 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:01.216 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:01.245 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:01.245 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:01 vm08.local ceph-mon[47734]: pgmap v2961: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:01 vm10.local ceph-mon[51861]: pgmap v2961: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:03 vm08.local ceph-mon[47734]: pgmap v2962: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:03 vm10.local ceph-mon[51861]: pgmap v2962: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:05.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:04 vm08.local ceph-mon[47734]: pgmap v2963: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:04 vm10.local ceph-mon[51861]: pgmap v2963: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:06.247 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:06.247 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:06.274 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:06.274 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:07.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:07 vm10.local ceph-mon[51861]: pgmap v2964: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:07 vm08.local ceph-mon[47734]: pgmap v2964: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:09 vm10.local ceph-mon[51861]: pgmap v2965: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:09 vm08.local ceph-mon[47734]: pgmap v2965: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:11.276 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:11.276 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:11.306 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:11.307 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:11 vm08.local ceph-mon[47734]: pgmap v2966: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:11 vm10.local ceph-mon[51861]: pgmap v2966: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:13 vm08.local ceph-mon[47734]: pgmap v2967: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:13 vm10.local ceph-mon[51861]: pgmap v2967: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:15 vm10.local ceph-mon[51861]: pgmap v2968: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:15 vm08.local ceph-mon[47734]: pgmap v2968: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:16.309 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:16.309 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:16.337 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:16.338 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:17.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:17 vm10.local ceph-mon[51861]: pgmap v2969: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:17 vm08.local ceph-mon[47734]: pgmap v2969: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:19 vm10.local ceph-mon[51861]: pgmap v2970: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:19 vm08.local ceph-mon[47734]: pgmap v2970: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:20 vm08.local ceph-mon[47734]: pgmap v2971: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:20 vm10.local ceph-mon[51861]: pgmap v2971: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:21.340 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:21.340 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:21.432 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:21.432 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:23 vm10.local ceph-mon[51861]: pgmap v2972: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:23 vm08.local ceph-mon[47734]: pgmap v2972: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:25.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:25 vm08.local ceph-mon[47734]: pgmap v2973: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:25 vm10.local ceph-mon[51861]: pgmap v2973: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:26.434 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:26.435 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:26.462 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:26.462 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:27.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:27 vm08.local ceph-mon[47734]: pgmap v2974: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:27 vm10.local ceph-mon[51861]: pgmap v2974: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:29.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:29 vm10.local ceph-mon[51861]: pgmap v2975: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:29 vm08.local ceph-mon[47734]: pgmap v2975: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:32:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:32:30.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:30 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:32:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:32:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:32:30.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:30 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:32:31.463 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:31.464 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:31.493 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:31.493 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:31 vm08.local ceph-mon[47734]: pgmap v2976: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:32:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:32:31.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:31 vm10.local ceph-mon[51861]: pgmap v2976: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:31.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:32:31.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:32:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:33 vm08.local ceph-mon[47734]: pgmap v2977: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:33 vm10.local ceph-mon[51861]: pgmap v2977: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:35.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:35 vm10.local ceph-mon[51861]: pgmap v2978: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:35 vm08.local ceph-mon[47734]: pgmap v2978: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:36.495 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:36.495 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:36.521 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:36.522 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:37.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:37 vm10.local ceph-mon[51861]: pgmap v2979: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:37 vm08.local ceph-mon[47734]: pgmap v2979: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:39.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:39 vm10.local ceph-mon[51861]: pgmap v2980: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:39 vm08.local ceph-mon[47734]: pgmap v2980: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:41.523 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:41.524 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:41.551 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:41.552 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:41 vm08.local ceph-mon[47734]: pgmap v2981: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:41 vm10.local ceph-mon[51861]: pgmap v2981: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:43 vm08.local ceph-mon[47734]: pgmap v2982: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:32:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:32:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:43 vm10.local ceph-mon[51861]: pgmap v2982: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:32:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:32:45.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:45 vm10.local ceph-mon[51861]: pgmap v2983: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:45 vm08.local ceph-mon[47734]: pgmap v2983: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:46.553 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:46.554 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:46.582 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:46.583 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:47 vm08.local ceph-mon[47734]: pgmap v2984: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:47 vm10.local ceph-mon[51861]: pgmap v2984: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:49 vm08.local ceph-mon[47734]: pgmap v2985: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:49 vm10.local ceph-mon[51861]: pgmap v2985: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:51 vm08.local ceph-mon[47734]: pgmap v2986: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:51.584 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:51.585 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:51.612 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:51.613 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:51 vm10.local ceph-mon[51861]: pgmap v2986: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:53 vm08.local ceph-mon[47734]: pgmap v2987: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:53 vm10.local ceph-mon[51861]: pgmap v2987: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:55 vm08.local ceph-mon[47734]: pgmap v2988: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:55 vm10.local ceph-mon[51861]: pgmap v2988: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:32:56.614 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:32:56.614 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:32:56.640 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:32:56.641 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:32:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:57 vm08.local ceph-mon[47734]: pgmap v2989: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:57 vm10.local ceph-mon[51861]: pgmap v2989: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:32:59 vm08.local ceph-mon[47734]: pgmap v2990: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:32:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:32:59 vm10.local ceph-mon[51861]: pgmap v2990: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:01 vm08.local ceph-mon[47734]: pgmap v2991: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:01.642 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:01.643 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:01 vm10.local ceph-mon[51861]: pgmap v2991: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:01.669 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:01.670 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:03 vm08.local ceph-mon[47734]: pgmap v2992: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:03 vm10.local ceph-mon[51861]: pgmap v2992: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:05.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:05 vm10.local ceph-mon[51861]: pgmap v2993: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:05 vm08.local ceph-mon[47734]: pgmap v2993: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:06.671 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:06.672 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:06.700 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:06.700 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:07 vm08.local ceph-mon[47734]: pgmap v2994: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:07 vm10.local ceph-mon[51861]: pgmap v2994: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:09.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:09 vm10.local ceph-mon[51861]: pgmap v2995: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:09 vm08.local ceph-mon[47734]: pgmap v2995: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:11 vm08.local ceph-mon[47734]: pgmap v2996: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:11 vm10.local ceph-mon[51861]: pgmap v2996: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:11.702 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:11.702 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:11.732 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:11.732 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:13 vm08.local ceph-mon[47734]: pgmap v2997: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:13 vm10.local ceph-mon[51861]: pgmap v2997: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:15.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:15 vm10.local ceph-mon[51861]: pgmap v2998: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:15 vm08.local ceph-mon[47734]: pgmap v2998: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:16.734 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:16.735 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:16.765 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:16.765 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:17.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:17 vm08.local ceph-mon[47734]: pgmap v2999: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:17 vm10.local ceph-mon[51861]: pgmap v2999: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:19.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:19 vm10.local ceph-mon[51861]: pgmap v3000: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:19 vm08.local ceph-mon[47734]: pgmap v3000: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:21.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:21 vm08.local ceph-mon[47734]: pgmap v3001: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:21 vm10.local ceph-mon[51861]: pgmap v3001: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:21.767 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:21.768 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:21.796 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:21.797 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:23 vm08.local ceph-mon[47734]: pgmap v3002: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:23 vm10.local ceph-mon[51861]: pgmap v3002: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:25.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:25 vm10.local ceph-mon[51861]: pgmap v3003: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:25 vm08.local ceph-mon[47734]: pgmap v3003: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:26.799 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:26.799 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:26.828 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:26.828 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:27 vm08.local ceph-mon[47734]: pgmap v3004: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:27 vm10.local ceph-mon[51861]: pgmap v3004: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:29.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:29 vm10.local ceph-mon[51861]: pgmap v3005: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:29.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:29 vm08.local ceph-mon[47734]: pgmap v3005: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:31.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:31 vm08.local ceph-mon[47734]: pgmap v3006: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:33:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:33:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:33:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:33:31.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:33:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:31 vm10.local ceph-mon[51861]: pgmap v3006: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:33:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:33:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:33:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:33:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:33:31.830 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:31.830 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:31.896 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:31.896 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:33 vm08.local ceph-mon[47734]: pgmap v3007: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:33 vm10.local ceph-mon[51861]: pgmap v3007: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:35 vm08.local ceph-mon[47734]: pgmap v3008: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:35 vm10.local ceph-mon[51861]: pgmap v3008: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:36.898 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:36.899 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:36.929 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:36.929 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:37.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:37 vm08.local ceph-mon[47734]: pgmap v3009: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:37 vm10.local ceph-mon[51861]: pgmap v3009: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:39 vm08.local ceph-mon[47734]: pgmap v3010: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:39 vm10.local ceph-mon[51861]: pgmap v3010: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:41.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:41 vm08.local ceph-mon[47734]: pgmap v3011: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:41 vm10.local ceph-mon[51861]: pgmap v3011: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:41.931 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:41.932 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:41.960 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:41.961 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:43 vm08.local ceph-mon[47734]: pgmap v3012: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:33:43.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:33:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:43 vm10.local ceph-mon[51861]: pgmap v3012: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:33:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:33:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:45 vm08.local ceph-mon[47734]: pgmap v3013: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:45 vm10.local ceph-mon[51861]: pgmap v3013: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:46 vm08.local ceph-mon[47734]: pgmap v3014: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:46.963 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:46.963 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:46.991 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:46.991 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:47.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:46 vm10.local ceph-mon[51861]: pgmap v3014: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:49 vm08.local ceph-mon[47734]: pgmap v3015: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:49 vm10.local ceph-mon[51861]: pgmap v3015: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:51.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:51 vm08.local ceph-mon[47734]: pgmap v3016: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:51 vm10.local ceph-mon[51861]: pgmap v3016: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:51.993 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:51.993 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:52.020 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:52.021 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:53 vm08.local ceph-mon[47734]: pgmap v3017: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:53 vm10.local ceph-mon[51861]: pgmap v3017: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:55 vm08.local ceph-mon[47734]: pgmap v3018: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:55 vm10.local ceph-mon[51861]: pgmap v3018: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:33:57.022 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:33:57.022 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:33:57.049 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:33:57.050 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:33:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:57 vm08.local ceph-mon[47734]: pgmap v3019: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:57 vm10.local ceph-mon[51861]: pgmap v3019: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:33:59 vm08.local ceph-mon[47734]: pgmap v3020: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:33:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:33:59 vm10.local ceph-mon[51861]: pgmap v3020: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:01 vm08.local ceph-mon[47734]: pgmap v3021: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:01 vm10.local ceph-mon[51861]: pgmap v3021: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:02.051 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:02.051 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:02.080 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:02.081 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:03 vm08.local ceph-mon[47734]: pgmap v3022: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:03 vm10.local ceph-mon[51861]: pgmap v3022: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:05 vm08.local ceph-mon[47734]: pgmap v3023: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:05 vm10.local ceph-mon[51861]: pgmap v3023: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:07.082 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:07.082 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:07.108 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:07.109 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:07.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:07 vm08.local ceph-mon[47734]: pgmap v3024: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:07 vm10.local ceph-mon[51861]: pgmap v3024: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:09 vm08.local ceph-mon[47734]: pgmap v3025: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:09 vm10.local ceph-mon[51861]: pgmap v3025: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:11 vm08.local ceph-mon[47734]: pgmap v3026: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:11 vm10.local ceph-mon[51861]: pgmap v3026: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:12.110 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:12.111 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:12.138 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:12.138 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:13 vm08.local ceph-mon[47734]: pgmap v3027: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:13 vm10.local ceph-mon[51861]: pgmap v3027: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:15 vm10.local ceph-mon[51861]: pgmap v3028: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:15.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:15 vm08.local ceph-mon[47734]: pgmap v3028: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:17.140 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:17.140 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:17.167 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:17.168 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:17 vm10.local ceph-mon[51861]: pgmap v3029: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:17 vm08.local ceph-mon[47734]: pgmap v3029: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:19 vm08.local ceph-mon[47734]: pgmap v3030: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:19 vm10.local ceph-mon[51861]: pgmap v3030: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:21 vm10.local ceph-mon[51861]: pgmap v3031: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:21.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:21 vm08.local ceph-mon[47734]: pgmap v3031: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:22.170 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:22.170 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:22.198 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:22.198 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:23 vm10.local ceph-mon[51861]: pgmap v3032: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:23 vm08.local ceph-mon[47734]: pgmap v3032: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:25 vm10.local ceph-mon[51861]: pgmap v3033: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:25.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:25 vm08.local ceph-mon[47734]: pgmap v3033: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:27.200 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:27.200 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:27.239 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:27.240 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:27 vm10.local ceph-mon[51861]: pgmap v3034: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:27 vm08.local ceph-mon[47734]: pgmap v3034: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:29 vm08.local ceph-mon[47734]: pgmap v3035: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:29 vm10.local ceph-mon[51861]: pgmap v3035: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:30.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:30 vm10.local ceph-mon[51861]: pgmap v3036: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:30.664 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:30 vm08.local ceph-mon[47734]: pgmap v3036: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:34:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:34:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:34:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:34:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:34:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:34:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:34:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:34:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:34:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:34:32.241 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:32.242 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:32.271 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:32.271 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:32 vm10.local ceph-mon[51861]: pgmap v3037: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:32 vm08.local ceph-mon[47734]: pgmap v3037: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:35 vm08.local ceph-mon[47734]: pgmap v3038: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:35 vm10.local ceph-mon[51861]: pgmap v3038: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:37.273 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:37.273 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:37.302 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:37.303 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:37 vm08.local ceph-mon[47734]: pgmap v3039: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:37 vm10.local ceph-mon[51861]: pgmap v3039: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:39 vm08.local ceph-mon[47734]: pgmap v3040: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:39 vm10.local ceph-mon[51861]: pgmap v3040: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:41 vm08.local ceph-mon[47734]: pgmap v3041: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:41 vm10.local ceph-mon[51861]: pgmap v3041: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:42.304 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:42.305 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:42.331 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:42.332 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:43 vm08.local ceph-mon[47734]: pgmap v3042: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:34:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:34:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:43 vm10.local ceph-mon[51861]: pgmap v3042: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:34:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:34:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:45 vm10.local ceph-mon[51861]: pgmap v3043: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:45 vm08.local ceph-mon[47734]: pgmap v3043: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:47.334 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:47.334 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:47.371 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:47.372 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:47 vm10.local ceph-mon[51861]: pgmap v3044: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:47 vm08.local ceph-mon[47734]: pgmap v3044: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:49 vm08.local ceph-mon[47734]: pgmap v3045: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:49 vm10.local ceph-mon[51861]: pgmap v3045: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:50 vm10.local ceph-mon[51861]: pgmap v3046: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:50 vm08.local ceph-mon[47734]: pgmap v3046: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:52.373 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:52.373 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:52.400 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:52.401 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:53 vm08.local ceph-mon[47734]: pgmap v3047: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:53 vm10.local ceph-mon[51861]: pgmap v3047: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:55 vm08.local ceph-mon[47734]: pgmap v3048: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:55 vm10.local ceph-mon[51861]: pgmap v3048: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:34:57.402 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:34:57.403 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:34:57.428 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:34:57.429 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:34:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:57 vm08.local ceph-mon[47734]: pgmap v3049: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:57 vm10.local ceph-mon[51861]: pgmap v3049: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:34:59 vm08.local ceph-mon[47734]: pgmap v3050: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:34:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:34:59 vm10.local ceph-mon[51861]: pgmap v3050: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:01 vm10.local ceph-mon[51861]: pgmap v3051: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:01.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:01 vm08.local ceph-mon[47734]: pgmap v3051: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:02.430 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:02.431 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:02.459 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:02.460 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:03 vm10.local ceph-mon[51861]: pgmap v3052: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:03 vm08.local ceph-mon[47734]: pgmap v3052: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:05 vm10.local ceph-mon[51861]: pgmap v3053: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:05.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:05 vm08.local ceph-mon[47734]: pgmap v3053: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:06.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:06 vm08.local ceph-mon[47734]: pgmap v3054: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:06 vm10.local ceph-mon[51861]: pgmap v3054: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:07.461 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:07.462 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:07.488 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:07.488 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:09 vm08.local ceph-mon[47734]: pgmap v3055: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:09 vm10.local ceph-mon[51861]: pgmap v3055: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:11 vm08.local ceph-mon[47734]: pgmap v3056: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:11 vm10.local ceph-mon[51861]: pgmap v3056: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:12.490 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:12.490 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:12.517 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:12.518 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:13 vm08.local ceph-mon[47734]: pgmap v3057: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:13 vm10.local ceph-mon[51861]: pgmap v3057: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:15 vm08.local ceph-mon[47734]: pgmap v3058: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:15 vm10.local ceph-mon[51861]: pgmap v3058: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:17.519 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:17.520 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:17.545 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:17.546 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:17 vm08.local ceph-mon[47734]: pgmap v3059: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:17 vm10.local ceph-mon[51861]: pgmap v3059: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:19 vm08.local ceph-mon[47734]: pgmap v3060: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:19 vm10.local ceph-mon[51861]: pgmap v3060: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:21.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:21 vm08.local ceph-mon[47734]: pgmap v3061: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:21 vm10.local ceph-mon[51861]: pgmap v3061: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:22.547 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:22.547 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:22.573 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:22.574 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:23 vm08.local ceph-mon[47734]: pgmap v3062: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:23 vm10.local ceph-mon[51861]: pgmap v3062: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:25.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:25 vm08.local ceph-mon[47734]: pgmap v3063: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:25 vm10.local ceph-mon[51861]: pgmap v3063: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:27.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:27 vm08.local ceph-mon[47734]: pgmap v3064: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:27.575 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:27.576 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:27.602 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:27.602 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:27 vm10.local ceph-mon[51861]: pgmap v3064: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:29 vm08.local ceph-mon[47734]: pgmap v3065: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:29 vm10.local ceph-mon[51861]: pgmap v3065: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:31.272 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:31 vm08.local ceph-mon[47734]: pgmap v3066: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:31.272 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:35:31.272 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:35:31.272 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:35:31.272 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:35:31.272 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:35:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:31 vm10.local ceph-mon[51861]: pgmap v3066: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:35:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:35:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:35:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:35:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:35:32.604 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:32.604 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:32.631 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:32.631 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:32 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:35:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:32 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:35:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:32 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:35:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:32 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:35:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:33 vm10.local ceph-mon[51861]: pgmap v3067: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:33 vm08.local ceph-mon[47734]: pgmap v3067: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:35 vm10.local ceph-mon[51861]: pgmap v3068: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:35 vm08.local ceph-mon[47734]: pgmap v3068: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:37.634 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:37.634 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:37 vm10.local ceph-mon[51861]: pgmap v3069: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:37.662 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:37.662 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:37.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:37 vm08.local ceph-mon[47734]: pgmap v3069: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:39 vm08.local ceph-mon[47734]: pgmap v3070: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:39 vm10.local ceph-mon[51861]: pgmap v3070: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:41 vm10.local ceph-mon[51861]: pgmap v3071: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:41.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:41 vm08.local ceph-mon[47734]: pgmap v3071: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:42.664 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:42.664 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:42.690 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:42.691 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:43 vm08.local ceph-mon[47734]: pgmap v3072: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:35:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:35:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:43 vm10.local ceph-mon[51861]: pgmap v3072: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:35:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:35:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:45 vm10.local ceph-mon[51861]: pgmap v3073: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:45 vm08.local ceph-mon[47734]: pgmap v3073: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:47.692 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:47.693 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:47.720 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:47.720 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:47 vm08.local ceph-mon[47734]: pgmap v3074: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:47.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:47 vm10.local ceph-mon[51861]: pgmap v3074: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:49.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:48 vm08.local ceph-mon[47734]: pgmap v3075: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:48 vm10.local ceph-mon[51861]: pgmap v3075: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:51.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:51 vm08.local ceph-mon[47734]: pgmap v3076: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:51.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:51 vm10.local ceph-mon[51861]: pgmap v3076: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:52.722 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:52.722 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:52.749 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:52.749 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:52 vm08.local ceph-mon[47734]: pgmap v3077: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:52 vm10.local ceph-mon[51861]: pgmap v3077: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:55 vm08.local ceph-mon[47734]: pgmap v3078: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:55 vm10.local ceph-mon[51861]: pgmap v3078: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:35:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:57 vm08.local ceph-mon[47734]: pgmap v3079: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:57 vm10.local ceph-mon[51861]: pgmap v3079: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:57.751 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:35:57.751 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:35:57.778 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:35:57.778 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:35:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:35:59 vm08.local ceph-mon[47734]: pgmap v3080: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:35:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:35:59 vm10.local ceph-mon[51861]: pgmap v3080: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:01.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:01 vm08.local ceph-mon[47734]: pgmap v3081: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:01 vm10.local ceph-mon[51861]: pgmap v3081: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:02.780 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:02.780 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:02.805 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:02.805 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:03 vm08.local ceph-mon[47734]: pgmap v3082: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:03 vm10.local ceph-mon[51861]: pgmap v3082: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:05 vm08.local ceph-mon[47734]: pgmap v3083: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:05 vm10.local ceph-mon[51861]: pgmap v3083: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:07 vm08.local ceph-mon[47734]: pgmap v3084: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:07 vm10.local ceph-mon[51861]: pgmap v3084: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:07.807 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:07.807 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:07.834 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:07.834 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:09 vm08.local ceph-mon[47734]: pgmap v3085: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:09 vm10.local ceph-mon[51861]: pgmap v3085: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:11 vm08.local ceph-mon[47734]: pgmap v3086: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:11 vm10.local ceph-mon[51861]: pgmap v3086: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:12.836 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:12.836 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:12.864 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:12.865 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:13 vm08.local ceph-mon[47734]: pgmap v3087: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:13 vm10.local ceph-mon[51861]: pgmap v3087: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:15 vm08.local ceph-mon[47734]: pgmap v3088: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:15 vm10.local ceph-mon[51861]: pgmap v3088: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:17 vm08.local ceph-mon[47734]: pgmap v3089: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:17 vm10.local ceph-mon[51861]: pgmap v3089: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:17.866 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:17.867 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:17.894 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:17.895 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:19 vm08.local ceph-mon[47734]: pgmap v3090: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:19 vm10.local ceph-mon[51861]: pgmap v3090: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:21 vm10.local ceph-mon[51861]: pgmap v3091: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:21.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:21 vm08.local ceph-mon[47734]: pgmap v3091: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:22.896 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:22.897 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:22.923 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:22.924 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:23 vm10.local ceph-mon[51861]: pgmap v3092: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:23 vm08.local ceph-mon[47734]: pgmap v3092: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:25 vm10.local ceph-mon[51861]: pgmap v3093: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:25 vm08.local ceph-mon[47734]: pgmap v3093: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:27 vm10.local ceph-mon[51861]: pgmap v3094: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:27 vm08.local ceph-mon[47734]: pgmap v3094: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:27.925 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:27.925 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:27.950 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:27.951 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:29 vm08.local ceph-mon[47734]: pgmap v3095: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:29 vm10.local ceph-mon[51861]: pgmap v3095: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:31 vm10.local ceph-mon[51861]: pgmap v3096: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:36:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:36:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:31 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:36:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:31 vm08.local ceph-mon[47734]: pgmap v3096: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:36:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:36:31.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:31 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:36:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:32 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:36:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:32 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:36:32.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:32 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:36:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:32 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:36:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:32 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:36:32.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:32 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:36:32.952 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:32.952 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:32.991 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:32.991 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:33 vm10.local ceph-mon[51861]: pgmap v3097: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:33 vm08.local ceph-mon[47734]: pgmap v3097: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:35 vm10.local ceph-mon[51861]: pgmap v3098: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:35 vm08.local ceph-mon[47734]: pgmap v3098: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:37 vm10.local ceph-mon[51861]: pgmap v3099: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:37.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:37 vm08.local ceph-mon[47734]: pgmap v3099: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:37.993 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:37.993 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:38.019 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:38.020 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:38 vm10.local ceph-mon[51861]: pgmap v3100: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:38 vm08.local ceph-mon[47734]: pgmap v3100: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:41.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:41 vm08.local ceph-mon[47734]: pgmap v3101: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:41.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:41 vm10.local ceph-mon[51861]: pgmap v3101: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:43.021 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:43.021 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:43.049 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:43.050 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:43 vm08.local ceph-mon[47734]: pgmap v3102: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:36:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:36:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:43 vm10.local ceph-mon[51861]: pgmap v3102: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:36:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:36:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:45 vm10.local ceph-mon[51861]: pgmap v3103: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:45 vm08.local ceph-mon[47734]: pgmap v3103: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:47 vm10.local ceph-mon[51861]: pgmap v3104: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:47 vm08.local ceph-mon[47734]: pgmap v3104: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:48.052 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:48.052 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:48.080 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:48.081 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:49 vm08.local ceph-mon[47734]: pgmap v3105: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:49 vm10.local ceph-mon[51861]: pgmap v3105: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:51 vm10.local ceph-mon[51861]: pgmap v3106: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:51.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:51 vm08.local ceph-mon[47734]: pgmap v3106: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:52.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:52 vm08.local ceph-mon[47734]: pgmap v3107: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:52.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:52 vm10.local ceph-mon[51861]: pgmap v3107: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:53.082 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:53.083 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:53.109 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:53.110 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:55 vm08.local ceph-mon[47734]: pgmap v3108: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:55 vm10.local ceph-mon[51861]: pgmap v3108: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:36:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:57 vm08.local ceph-mon[47734]: pgmap v3109: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:57 vm10.local ceph-mon[51861]: pgmap v3109: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:58.111 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:36:58.111 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:36:58.137 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:36:58.138 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:36:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:36:59 vm08.local ceph-mon[47734]: pgmap v3110: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:36:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:36:59 vm10.local ceph-mon[51861]: pgmap v3110: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:01.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:01 vm10.local ceph-mon[51861]: pgmap v3111: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:01.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:01 vm08.local ceph-mon[47734]: pgmap v3111: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:03.139 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:03.139 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:03.166 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:03.167 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:03 vm10.local ceph-mon[51861]: pgmap v3112: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:03 vm08.local ceph-mon[47734]: pgmap v3112: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:05 vm10.local ceph-mon[51861]: pgmap v3113: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:05 vm08.local ceph-mon[47734]: pgmap v3113: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:06.708 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:06 vm08.local ceph-mon[47734]: pgmap v3114: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:06.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:06 vm10.local ceph-mon[51861]: pgmap v3114: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:08.168 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:08.169 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:08.195 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:08.195 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:09 vm08.local ceph-mon[47734]: pgmap v3115: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:09 vm10.local ceph-mon[51861]: pgmap v3115: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:11.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:11 vm08.local ceph-mon[47734]: pgmap v3116: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:11 vm10.local ceph-mon[51861]: pgmap v3116: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:13.196 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:13.197 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:13.226 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:13.226 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:13 vm10.local ceph-mon[51861]: pgmap v3117: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:13.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:13 vm08.local ceph-mon[47734]: pgmap v3117: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:15 vm10.local ceph-mon[51861]: pgmap v3118: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:15 vm08.local ceph-mon[47734]: pgmap v3118: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:17 vm10.local ceph-mon[51861]: pgmap v3119: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:17 vm08.local ceph-mon[47734]: pgmap v3119: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:18.228 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:18.229 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:18.256 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:18.257 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:19 vm08.local ceph-mon[47734]: pgmap v3120: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:19 vm10.local ceph-mon[51861]: pgmap v3120: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:21 vm10.local ceph-mon[51861]: pgmap v3121: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:21.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:21 vm08.local ceph-mon[47734]: pgmap v3121: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:23.258 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:23.258 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:23.286 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:23.287 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:23 vm10.local ceph-mon[51861]: pgmap v3122: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:23 vm08.local ceph-mon[47734]: pgmap v3122: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:25 vm10.local ceph-mon[51861]: pgmap v3123: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:25 vm08.local ceph-mon[47734]: pgmap v3123: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:27 vm10.local ceph-mon[51861]: pgmap v3124: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:27 vm08.local ceph-mon[47734]: pgmap v3124: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:28.288 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:28.289 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:28.317 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:28.317 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:29 vm08.local ceph-mon[47734]: pgmap v3125: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:29 vm10.local ceph-mon[51861]: pgmap v3125: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:31.634 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:31 vm08.local ceph-mon[47734]: pgmap v3126: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:31.635 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:31 vm10.local ceph-mon[51861]: pgmap v3126: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:32.374 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:32 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:37:32.374 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:32 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:37:32.374 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:32 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:37:32.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:32 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:37:32.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:32 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:37:32.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:32 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:37:33.318 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:33.319 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:33.356 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:33.356 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:33.377 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:33 vm08.local ceph-mon[47734]: pgmap v3127: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:33.377 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:33 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:37:33.377 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:33 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:37:33.377 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:33 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:37:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:33 vm10.local ceph-mon[51861]: pgmap v3127: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:33 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:37:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:33 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:37:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:33 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:37:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:37:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:37:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:34 vm10.local ceph-mon[51861]: pgmap v3128: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:37:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:37:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:34 vm08.local ceph-mon[47734]: pgmap v3128: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:37 vm08.local ceph-mon[47734]: pgmap v3129: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:37 vm10.local ceph-mon[51861]: pgmap v3129: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:38.357 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:38.358 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:38.385 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:38.385 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:39 vm08.local ceph-mon[47734]: pgmap v3130: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:39 vm10.local ceph-mon[51861]: pgmap v3130: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:40 vm08.local ceph-mon[47734]: pgmap v3131: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:40 vm10.local ceph-mon[51861]: pgmap v3131: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:43.387 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:43.387 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:43.416 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:43.417 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:43 vm10.local ceph-mon[51861]: pgmap v3132: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:37:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:37:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:43 vm08.local ceph-mon[47734]: pgmap v3132: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:37:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:37:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:45 vm08.local ceph-mon[47734]: pgmap v3133: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:45 vm10.local ceph-mon[51861]: pgmap v3133: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:46.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:46 vm08.local ceph-mon[47734]: pgmap v3134: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:46 vm10.local ceph-mon[51861]: pgmap v3134: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:48.418 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:48.419 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:48.447 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:48.447 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:49.483 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:49 vm10.local ceph-mon[51861]: pgmap v3135: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:49 vm08.local ceph-mon[47734]: pgmap v3135: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:51 vm10.local ceph-mon[51861]: pgmap v3136: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:51.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:51 vm08.local ceph-mon[47734]: pgmap v3136: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:53.449 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:53.449 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:53.476 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:53.477 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:53 vm10.local ceph-mon[51861]: pgmap v3137: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:53 vm08.local ceph-mon[47734]: pgmap v3137: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:55 vm10.local ceph-mon[51861]: pgmap v3138: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:55 vm08.local ceph-mon[47734]: pgmap v3138: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:37:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:57 vm10.local ceph-mon[51861]: pgmap v3139: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:57 vm08.local ceph-mon[47734]: pgmap v3139: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:58.478 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:37:58.478 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:37:58.503 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:37:58.504 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:37:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:37:59 vm08.local ceph-mon[47734]: pgmap v3140: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:37:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:37:59 vm10.local ceph-mon[51861]: pgmap v3140: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:00 vm08.local ceph-mon[47734]: pgmap v3141: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:00 vm10.local ceph-mon[51861]: pgmap v3141: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:03.505 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:03.506 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:03.534 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:03.534 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:03 vm08.local ceph-mon[47734]: pgmap v3142: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:03 vm10.local ceph-mon[51861]: pgmap v3142: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:05 vm10.local ceph-mon[51861]: pgmap v3143: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:05 vm08.local ceph-mon[47734]: pgmap v3143: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:07 vm10.local ceph-mon[51861]: pgmap v3144: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:07 vm08.local ceph-mon[47734]: pgmap v3144: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:08.535 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:08.536 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:08.561 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:08.562 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:09 vm08.local ceph-mon[47734]: pgmap v3145: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:09 vm10.local ceph-mon[51861]: pgmap v3145: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:11 vm10.local ceph-mon[51861]: pgmap v3146: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:11.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:11 vm08.local ceph-mon[47734]: pgmap v3146: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:13.563 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:13.563 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:13.590 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:13.590 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:13 vm10.local ceph-mon[51861]: pgmap v3147: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:13.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:13 vm08.local ceph-mon[47734]: pgmap v3147: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:15 vm10.local ceph-mon[51861]: pgmap v3148: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:15 vm08.local ceph-mon[47734]: pgmap v3148: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:17 vm10.local ceph-mon[51861]: pgmap v3149: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:17 vm08.local ceph-mon[47734]: pgmap v3149: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:18.592 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:18.592 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:18.619 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:18.620 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:19 vm08.local ceph-mon[47734]: pgmap v3150: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:19 vm10.local ceph-mon[51861]: pgmap v3150: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:21 vm10.local ceph-mon[51861]: pgmap v3151: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:21.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:21 vm08.local ceph-mon[47734]: pgmap v3151: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:23.621 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:23.622 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:23.648 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:23.648 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:23 vm10.local ceph-mon[51861]: pgmap v3152: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:23 vm08.local ceph-mon[47734]: pgmap v3152: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:25 vm10.local ceph-mon[51861]: pgmap v3153: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:25 vm08.local ceph-mon[47734]: pgmap v3153: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:27 vm10.local ceph-mon[51861]: pgmap v3154: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:27 vm08.local ceph-mon[47734]: pgmap v3154: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:28.649 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:28.650 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:28.676 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:28.676 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:29 vm08.local ceph-mon[47734]: pgmap v3155: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:29 vm10.local ceph-mon[51861]: pgmap v3155: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:31 vm10.local ceph-mon[51861]: pgmap v3156: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:31.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:31 vm08.local ceph-mon[47734]: pgmap v3156: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:33.646 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:33 vm08.local ceph-mon[47734]: pgmap v3157: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:33 vm10.local ceph-mon[51861]: pgmap v3157: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:33.678 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:33.678 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:33.708 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:33.708 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:38:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:38:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:38:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:38:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:38:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:38:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:38:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:38:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:38:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:38:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:35 vm10.local ceph-mon[51861]: pgmap v3158: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:35 vm08.local ceph-mon[47734]: pgmap v3158: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:37 vm10.local ceph-mon[51861]: pgmap v3159: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:37.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:37 vm08.local ceph-mon[47734]: pgmap v3159: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:38.709 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:38.710 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:38.738 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:38.739 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:39 vm10.local ceph-mon[51861]: pgmap v3160: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:39.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:39 vm08.local ceph-mon[47734]: pgmap v3160: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:40 vm08.local ceph-mon[47734]: pgmap v3161: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:40 vm10.local ceph-mon[51861]: pgmap v3161: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:43 vm08.local ceph-mon[47734]: pgmap v3162: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:38:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:38:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:38:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:43 vm10.local ceph-mon[51861]: pgmap v3162: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:38:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:38:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:38:43.740 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:43.741 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:43.767 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:43.768 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:45 vm10.local ceph-mon[51861]: pgmap v3163: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:45 vm08.local ceph-mon[47734]: pgmap v3163: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:47.417 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:47 vm08.local ceph-mon[47734]: pgmap v3164: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:38:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:47 vm10.local ceph-mon[51861]: pgmap v3164: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:38:48.769 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:48.770 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:48.795 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:48.796 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:49 vm08.local ceph-mon[47734]: pgmap v3165: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:38:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:49 vm10.local ceph-mon[51861]: pgmap v3165: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:38:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:51 vm10.local ceph-mon[51861]: pgmap v3166: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:51.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:51 vm08.local ceph-mon[47734]: pgmap v3166: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:53 vm10.local ceph-mon[51861]: pgmap v3167: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:53.797 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:53.798 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:53 vm08.local ceph-mon[47734]: pgmap v3167: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:53.825 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:53.825 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:55 vm10.local ceph-mon[51861]: pgmap v3168: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:55 vm08.local ceph-mon[47734]: pgmap v3168: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:38:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:57 vm10.local ceph-mon[51861]: pgmap v3169: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:57.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:57 vm08.local ceph-mon[47734]: pgmap v3169: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:58.827 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:38:58.827 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:38:58.856 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:38:58.856 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:38:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:38:59 vm08.local ceph-mon[47734]: pgmap v3170: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:38:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:38:59 vm10.local ceph-mon[51861]: pgmap v3170: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:00 vm08.local ceph-mon[47734]: pgmap v3171: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:00 vm10.local ceph-mon[51861]: pgmap v3171: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:03 vm08.local ceph-mon[47734]: pgmap v3172: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:03 vm10.local ceph-mon[51861]: pgmap v3172: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:03.857 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:03.858 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:03.885 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:03.885 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:05 vm10.local ceph-mon[51861]: pgmap v3173: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:05 vm08.local ceph-mon[47734]: pgmap v3173: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:07 vm10.local ceph-mon[51861]: pgmap v3174: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:07 vm08.local ceph-mon[47734]: pgmap v3174: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:08.887 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:08.887 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:08.913 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:08.914 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:09 vm08.local ceph-mon[47734]: pgmap v3175: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:09 vm10.local ceph-mon[51861]: pgmap v3175: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:10 vm08.local ceph-mon[47734]: pgmap v3176: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:10 vm10.local ceph-mon[51861]: pgmap v3176: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:13 vm08.local ceph-mon[47734]: pgmap v3177: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:13 vm10.local ceph-mon[51861]: pgmap v3177: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:13.915 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:13.915 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:13.943 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:13.944 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:15 vm08.local ceph-mon[47734]: pgmap v3178: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:15 vm10.local ceph-mon[51861]: pgmap v3178: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:17.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:17 vm08.local ceph-mon[47734]: pgmap v3179: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:17 vm10.local ceph-mon[51861]: pgmap v3179: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:18.945 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:18.946 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:18.973 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:18.973 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:19 vm08.local ceph-mon[47734]: pgmap v3180: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:19 vm10.local ceph-mon[51861]: pgmap v3180: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:21 vm10.local ceph-mon[51861]: pgmap v3181: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:21.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:21 vm08.local ceph-mon[47734]: pgmap v3181: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:23 vm10.local ceph-mon[51861]: pgmap v3182: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:23 vm08.local ceph-mon[47734]: pgmap v3182: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:23.975 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:23.975 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:24.002 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:24.003 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:25 vm10.local ceph-mon[51861]: pgmap v3183: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:25 vm08.local ceph-mon[47734]: pgmap v3183: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:27 vm10.local ceph-mon[51861]: pgmap v3184: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:27 vm08.local ceph-mon[47734]: pgmap v3184: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:29.004 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:29.005 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:29.030 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:29.031 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:29 vm08.local ceph-mon[47734]: pgmap v3185: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:29 vm10.local ceph-mon[51861]: pgmap v3185: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:31.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:31 vm10.local ceph-mon[51861]: pgmap v3186: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:31.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:31 vm08.local ceph-mon[47734]: pgmap v3186: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:33 vm10.local ceph-mon[51861]: pgmap v3187: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:33.687 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:33 vm08.local ceph-mon[47734]: pgmap v3187: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:34.032 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:34.033 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:34.073 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:34.074 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:39:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:39:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:39:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:39:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:39:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:39:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:39:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:39:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:39:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:39:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:35 vm10.local ceph-mon[51861]: pgmap v3188: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:35 vm08.local ceph-mon[47734]: pgmap v3188: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:37.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:37 vm08.local ceph-mon[47734]: pgmap v3189: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:37.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:37 vm10.local ceph-mon[51861]: pgmap v3189: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:38 vm08.local ceph-mon[47734]: pgmap v3190: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:38.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:38 vm10.local ceph-mon[51861]: pgmap v3190: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:39.075 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:39.076 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:39.102 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:39.103 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:41.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:41 vm08.local ceph-mon[47734]: pgmap v3191: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:41.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:41 vm10.local ceph-mon[51861]: pgmap v3191: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:42.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:42 vm08.local ceph-mon[47734]: pgmap v3192: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:42.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:42 vm10.local ceph-mon[51861]: pgmap v3192: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:39:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:39:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:39:43.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:39:44.105 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:44.106 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:44.136 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:44.136 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:45 vm10.local ceph-mon[51861]: pgmap v3193: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:45 vm08.local ceph-mon[47734]: pgmap v3193: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:47 vm10.local ceph-mon[51861]: pgmap v3194: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:47 vm08.local ceph-mon[47734]: pgmap v3194: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:49.138 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:49.138 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:49.164 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:49.164 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:49 vm08.local ceph-mon[47734]: pgmap v3195: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:49 vm10.local ceph-mon[51861]: pgmap v3195: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:51.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:51 vm10.local ceph-mon[51861]: pgmap v3196: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:51.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:51 vm08.local ceph-mon[47734]: pgmap v3196: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:53 vm10.local ceph-mon[51861]: pgmap v3197: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:53 vm08.local ceph-mon[47734]: pgmap v3197: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:54.165 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:54.166 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:54.194 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:54.195 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:55 vm10.local ceph-mon[51861]: pgmap v3198: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:55 vm08.local ceph-mon[47734]: pgmap v3198: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:39:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:57 vm08.local ceph-mon[47734]: pgmap v3199: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:57.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:57 vm10.local ceph-mon[51861]: pgmap v3199: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:59.196 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:39:59.197 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:39:59.223 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:39:59.223 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:39:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:39:59 vm08.local ceph-mon[47734]: pgmap v3200: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:39:59.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:39:59 vm10.local ceph-mon[51861]: pgmap v3200: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T15:40:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T15:40:01.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:01 vm08.local ceph-mon[47734]: pgmap v3201: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:01.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:01 vm10.local ceph-mon[51861]: pgmap v3201: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:02.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:02 vm08.local ceph-mon[47734]: pgmap v3202: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:02.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:02 vm10.local ceph-mon[51861]: pgmap v3202: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:04.225 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:04.225 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:04.250 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:04.250 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:05 vm10.local ceph-mon[51861]: pgmap v3203: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:05 vm08.local ceph-mon[47734]: pgmap v3203: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:07 vm10.local ceph-mon[51861]: pgmap v3204: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:07.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:07 vm08.local ceph-mon[47734]: pgmap v3204: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:09.252 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:09.252 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:09.281 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:09.281 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:09 vm08.local ceph-mon[47734]: pgmap v3205: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:09 vm10.local ceph-mon[51861]: pgmap v3205: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:11.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:11 vm10.local ceph-mon[51861]: pgmap v3206: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:11.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:11 vm08.local ceph-mon[47734]: pgmap v3206: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:12.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:12 vm08.local ceph-mon[47734]: pgmap v3207: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:12.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:12 vm10.local ceph-mon[51861]: pgmap v3207: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:14.282 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:14.283 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:14.311 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:14.312 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:16.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:15 vm08.local ceph-mon[47734]: pgmap v3208: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:16.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:15 vm10.local ceph-mon[51861]: pgmap v3208: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:17.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:16 vm10.local ceph-mon[51861]: pgmap v3209: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:17.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:16 vm08.local ceph-mon[47734]: pgmap v3209: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:19.313 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:19.313 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:19.342 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:19.342 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:19 vm08.local ceph-mon[47734]: pgmap v3210: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:19 vm10.local ceph-mon[51861]: pgmap v3210: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:21.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:21 vm10.local ceph-mon[51861]: pgmap v3211: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:21.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:21 vm08.local ceph-mon[47734]: pgmap v3211: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:23 vm10.local ceph-mon[51861]: pgmap v3212: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:23 vm08.local ceph-mon[47734]: pgmap v3212: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:24.344 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:24.344 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:24.371 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:24.372 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:25 vm10.local ceph-mon[51861]: pgmap v3213: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:25 vm08.local ceph-mon[47734]: pgmap v3213: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:27.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:27 vm08.local ceph-mon[47734]: pgmap v3214: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:27.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:27 vm10.local ceph-mon[51861]: pgmap v3214: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:29.373 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:29.374 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:29.432 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:29.432 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:29.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:29 vm08.local ceph-mon[47734]: pgmap v3215: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:29.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:29 vm10.local ceph-mon[51861]: pgmap v3215: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:30.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:30 vm08.local ceph-mon[47734]: pgmap v3216: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:30 vm10.local ceph-mon[51861]: pgmap v3216: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:33 vm08.local ceph-mon[47734]: pgmap v3217: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:33 vm10.local ceph-mon[51861]: pgmap v3217: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:34.285 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:40:34.285 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:40:34.434 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:34.435 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:34.462 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:34.463 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:40:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:40:34.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:34 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:40:34.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:34 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:40:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:35 vm08.local ceph-mon[47734]: pgmap v3218: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:40:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:40:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:35 vm10.local ceph-mon[51861]: pgmap v3218: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:40:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:40:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:37 vm08.local ceph-mon[47734]: pgmap v3219: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:37 vm10.local ceph-mon[51861]: pgmap v3219: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:39.464 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:39.483 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:39 vm08.local ceph-mon[47734]: pgmap v3220: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:39.561 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:39.561 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:39 vm10.local ceph-mon[51861]: pgmap v3220: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:40 vm08.local ceph-mon[47734]: pgmap v3221: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:40 vm10.local ceph-mon[51861]: pgmap v3221: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:43 vm08.local ceph-mon[47734]: pgmap v3222: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:43 vm10.local ceph-mon[51861]: pgmap v3222: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:44.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:40:44.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:40:44.564 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:44.565 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:44.594 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:44.595 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:40:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:40:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:45 vm08.local ceph-mon[47734]: pgmap v3223: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:45 vm10.local ceph-mon[51861]: pgmap v3223: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:47 vm08.local ceph-mon[47734]: pgmap v3224: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:47 vm10.local ceph-mon[51861]: pgmap v3224: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:49 vm08.local ceph-mon[47734]: pgmap v3225: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:49.596 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:49.596 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:49.626 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:49.627 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:49 vm10.local ceph-mon[51861]: pgmap v3225: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:50 vm08.local ceph-mon[47734]: pgmap v3226: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:50 vm10.local ceph-mon[51861]: pgmap v3226: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:53 vm08.local ceph-mon[47734]: pgmap v3227: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:53 vm10.local ceph-mon[51861]: pgmap v3227: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:54.629 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:54.629 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:54.701 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:54.701 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:40:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:55 vm08.local ceph-mon[47734]: pgmap v3228: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:55 vm10.local ceph-mon[51861]: pgmap v3228: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:40:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:57 vm10.local ceph-mon[51861]: pgmap v3229: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:57 vm08.local ceph-mon[47734]: pgmap v3229: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:40:58 vm08.local ceph-mon[47734]: pgmap v3230: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:40:58 vm10.local ceph-mon[51861]: pgmap v3230: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:40:59.703 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:40:59.703 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:40:59.730 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:40:59.731 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:00.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:00 vm08.local ceph-mon[47734]: pgmap v3231: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:00 vm10.local ceph-mon[51861]: pgmap v3231: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:03 vm08.local ceph-mon[47734]: pgmap v3232: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:03 vm10.local ceph-mon[51861]: pgmap v3232: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:04.733 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:04.733 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:04.762 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:04.763 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:05 vm10.local ceph-mon[51861]: pgmap v3233: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:05 vm08.local ceph-mon[47734]: pgmap v3233: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:07 vm10.local ceph-mon[51861]: pgmap v3234: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:07.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:07 vm08.local ceph-mon[47734]: pgmap v3234: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:09 vm08.local ceph-mon[47734]: pgmap v3235: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:09 vm10.local ceph-mon[51861]: pgmap v3235: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:09.764 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:09.765 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:09.792 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:09.793 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:10.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:10 vm08.local ceph-mon[47734]: pgmap v3236: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:10 vm10.local ceph-mon[51861]: pgmap v3236: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:13 vm08.local ceph-mon[47734]: pgmap v3237: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:13 vm10.local ceph-mon[51861]: pgmap v3237: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:14.795 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:14.795 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:14.821 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:14.822 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:15 vm10.local ceph-mon[51861]: pgmap v3238: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:15 vm08.local ceph-mon[47734]: pgmap v3238: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:17 vm08.local ceph-mon[47734]: pgmap v3239: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:17.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:17 vm10.local ceph-mon[51861]: pgmap v3239: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:19.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:19 vm08.local ceph-mon[47734]: pgmap v3240: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:19.824 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:19.824 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:19.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:19 vm10.local ceph-mon[51861]: pgmap v3240: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:19.951 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:19.951 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:20.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:20 vm08.local ceph-mon[47734]: pgmap v3241: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:20 vm10.local ceph-mon[51861]: pgmap v3241: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:23 vm10.local ceph-mon[51861]: pgmap v3242: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:23 vm08.local ceph-mon[47734]: pgmap v3242: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:24.952 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:24.953 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:24.980 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:24.980 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:25 vm10.local ceph-mon[51861]: pgmap v3243: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:25 vm08.local ceph-mon[47734]: pgmap v3243: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:27 vm10.local ceph-mon[51861]: pgmap v3244: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:27 vm08.local ceph-mon[47734]: pgmap v3244: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:29 vm08.local ceph-mon[47734]: pgmap v3245: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:29 vm10.local ceph-mon[51861]: pgmap v3245: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:29.982 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:29.982 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:30.011 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:30.012 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:30 vm10.local ceph-mon[51861]: pgmap v3246: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:30 vm08.local ceph-mon[47734]: pgmap v3246: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:33 vm08.local ceph-mon[47734]: pgmap v3247: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:33 vm10.local ceph-mon[51861]: pgmap v3247: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:35.013 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:35.014 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:35.041 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:35.041 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:35 vm08.local ceph-mon[47734]: pgmap v3248: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:41:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:41:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:41:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:41:35.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:41:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:35 vm10.local ceph-mon[51861]: pgmap v3248: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:41:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:41:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:41:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:41:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:41:37.508 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:37 vm08.local ceph-mon[47734]: pgmap v3249: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:37 vm10.local ceph-mon[51861]: pgmap v3249: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:39 vm08.local ceph-mon[47734]: pgmap v3250: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:39 vm10.local ceph-mon[51861]: pgmap v3250: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:40.043 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:40.044 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:40.071 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:40.072 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:40.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:40 vm08.local ceph-mon[47734]: pgmap v3251: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:40 vm10.local ceph-mon[51861]: pgmap v3251: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:43 vm08.local ceph-mon[47734]: pgmap v3252: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:43 vm10.local ceph-mon[51861]: pgmap v3252: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:41:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:41:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:41:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:41:45.073 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:45.074 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:45.103 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:45.103 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:45 vm08.local ceph-mon[47734]: pgmap v3253: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:45 vm10.local ceph-mon[51861]: pgmap v3253: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:47.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:47 vm08.local ceph-mon[47734]: pgmap v3254: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:47 vm10.local ceph-mon[51861]: pgmap v3254: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:49 vm08.local ceph-mon[47734]: pgmap v3255: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:49 vm10.local ceph-mon[51861]: pgmap v3255: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:50.104 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:50.105 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:50.131 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:50.132 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:50.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:50 vm08.local ceph-mon[47734]: pgmap v3256: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:50 vm10.local ceph-mon[51861]: pgmap v3256: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:53.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:53 vm08.local ceph-mon[47734]: pgmap v3257: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:53 vm10.local ceph-mon[51861]: pgmap v3257: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:55.133 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:41:55.134 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:41:55.160 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:41:55.161 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:41:55.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:55 vm08.local ceph-mon[47734]: pgmap v3258: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:55 vm10.local ceph-mon[51861]: pgmap v3258: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:41:57.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:57 vm08.local ceph-mon[47734]: pgmap v3259: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:57 vm10.local ceph-mon[51861]: pgmap v3259: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:41:59 vm08.local ceph-mon[47734]: pgmap v3260: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:41:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:41:59 vm10.local ceph-mon[51861]: pgmap v3260: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:00.162 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:00.163 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:00.189 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:00.190 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:00 vm10.local ceph-mon[51861]: pgmap v3261: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:00 vm08.local ceph-mon[47734]: pgmap v3261: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:03 vm08.local ceph-mon[47734]: pgmap v3262: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:03 vm10.local ceph-mon[51861]: pgmap v3262: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:05.191 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:05.192 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:05.219 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:05.220 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:05.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:05 vm08.local ceph-mon[47734]: pgmap v3263: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:05 vm10.local ceph-mon[51861]: pgmap v3263: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:07.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:07 vm08.local ceph-mon[47734]: pgmap v3264: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:07 vm10.local ceph-mon[51861]: pgmap v3264: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:09 vm08.local ceph-mon[47734]: pgmap v3265: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:09 vm10.local ceph-mon[51861]: pgmap v3265: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:10.222 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:10.222 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:10.253 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:10.254 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:10 vm10.local ceph-mon[51861]: pgmap v3266: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:10 vm08.local ceph-mon[47734]: pgmap v3266: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:13 vm08.local ceph-mon[47734]: pgmap v3267: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:13 vm10.local ceph-mon[51861]: pgmap v3267: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:15.255 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:15.256 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:15.285 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:15.285 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:15.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:15 vm08.local ceph-mon[47734]: pgmap v3268: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:15 vm10.local ceph-mon[51861]: pgmap v3268: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:16.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:16 vm08.local ceph-mon[47734]: pgmap v3269: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:16.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:16 vm10.local ceph-mon[51861]: pgmap v3269: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:19 vm08.local ceph-mon[47734]: pgmap v3270: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:19.562 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:19 vm10.local ceph-mon[51861]: pgmap v3270: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:20.287 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:20.287 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:20.315 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:20.315 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:20 vm10.local ceph-mon[51861]: pgmap v3271: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:20 vm08.local ceph-mon[47734]: pgmap v3271: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:23 vm08.local ceph-mon[47734]: pgmap v3272: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:23 vm10.local ceph-mon[51861]: pgmap v3272: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:25.317 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:25.317 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:25.344 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:25.344 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:25.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:25 vm08.local ceph-mon[47734]: pgmap v3273: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:25 vm10.local ceph-mon[51861]: pgmap v3273: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:27 vm10.local ceph-mon[51861]: pgmap v3274: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:27 vm08.local ceph-mon[47734]: pgmap v3274: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:29 vm08.local ceph-mon[47734]: pgmap v3275: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:29 vm10.local ceph-mon[51861]: pgmap v3275: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:30.345 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:30.346 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:30.372 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:30.373 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:30 vm10.local ceph-mon[51861]: pgmap v3276: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:30 vm08.local ceph-mon[47734]: pgmap v3276: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:33.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:33 vm08.local ceph-mon[47734]: pgmap v3277: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:33 vm10.local ceph-mon[51861]: pgmap v3277: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:35.374 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:35.375 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:35.401 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:35.402 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:35.558 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:35 vm08.local ceph-mon[47734]: pgmap v3278: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:35.558 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:42:35.558 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:42:35.558 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:42:35.558 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:42:35.558 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:42:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:35 vm10.local ceph-mon[51861]: pgmap v3278: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:42:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:42:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:42:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:42:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:42:37.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:37 vm08.local ceph-mon[47734]: pgmap v3279: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:37 vm10.local ceph-mon[51861]: pgmap v3279: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:39 vm08.local ceph-mon[47734]: pgmap v3280: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:39 vm10.local ceph-mon[51861]: pgmap v3280: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:40.404 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:40.404 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:40.433 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:40.434 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:40.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:40 vm10.local ceph-mon[51861]: pgmap v3281: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:40 vm08.local ceph-mon[47734]: pgmap v3281: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:43 vm08.local ceph-mon[47734]: pgmap v3282: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:43 vm10.local ceph-mon[51861]: pgmap v3282: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:42:44.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:42:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:42:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:42:45.436 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:45.436 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:45.462 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:45.463 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:45.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:45 vm08.local ceph-mon[47734]: pgmap v3283: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:45 vm10.local ceph-mon[51861]: pgmap v3283: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:47 vm10.local ceph-mon[51861]: pgmap v3284: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:47 vm08.local ceph-mon[47734]: pgmap v3284: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:49 vm08.local ceph-mon[47734]: pgmap v3285: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:49 vm10.local ceph-mon[51861]: pgmap v3285: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:50.464 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:50.464 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:50.491 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:50.492 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:50 vm10.local ceph-mon[51861]: pgmap v3286: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:50 vm08.local ceph-mon[47734]: pgmap v3286: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:53.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:53 vm08.local ceph-mon[47734]: pgmap v3287: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:53.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:53 vm10.local ceph-mon[51861]: pgmap v3287: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:55.493 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:42:55.494 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:42:55.521 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:42:55.521 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:42:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:55 vm08.local ceph-mon[47734]: pgmap v3288: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:55 vm10.local ceph-mon[51861]: pgmap v3288: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:42:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:57 vm08.local ceph-mon[47734]: pgmap v3289: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:57.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:57 vm10.local ceph-mon[51861]: pgmap v3289: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:59.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:42:58 vm08.local ceph-mon[47734]: pgmap v3290: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:42:59.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:42:58 vm10.local ceph-mon[51861]: pgmap v3290: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:00.523 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:00.523 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:00.737 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:00.738 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:00 vm08.local ceph-mon[47734]: pgmap v3291: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:00 vm10.local ceph-mon[51861]: pgmap v3291: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:03.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:03 vm08.local ceph-mon[47734]: pgmap v3292: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:03 vm10.local ceph-mon[51861]: pgmap v3292: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:05 vm10.local ceph-mon[51861]: pgmap v3293: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:05.739 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:05.740 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:05 vm08.local ceph-mon[47734]: pgmap v3293: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:05.828 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:05.828 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:07 vm10.local ceph-mon[51861]: pgmap v3294: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:07 vm08.local ceph-mon[47734]: pgmap v3294: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:09 vm08.local ceph-mon[47734]: pgmap v3295: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:09 vm10.local ceph-mon[51861]: pgmap v3295: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:10.829 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:10.830 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:10 vm10.local ceph-mon[51861]: pgmap v3296: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:10.932 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:10.933 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:10 vm08.local ceph-mon[47734]: pgmap v3296: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:13.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:13 vm08.local ceph-mon[47734]: pgmap v3297: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:13 vm10.local ceph-mon[51861]: pgmap v3297: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:15 vm10.local ceph-mon[51861]: pgmap v3298: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:15 vm08.local ceph-mon[47734]: pgmap v3298: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:15.934 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:15.935 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:15.962 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:15.962 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:17 vm10.local ceph-mon[51861]: pgmap v3299: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:17 vm08.local ceph-mon[47734]: pgmap v3299: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:19 vm08.local ceph-mon[47734]: pgmap v3300: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:19 vm10.local ceph-mon[51861]: pgmap v3300: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:20.964 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:20.964 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:21.053 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:21.054 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:20 vm08.local ceph-mon[47734]: pgmap v3301: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:20 vm10.local ceph-mon[51861]: pgmap v3301: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:23.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:23 vm08.local ceph-mon[47734]: pgmap v3302: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:23 vm10.local ceph-mon[51861]: pgmap v3302: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:25 vm10.local ceph-mon[51861]: pgmap v3303: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:25 vm08.local ceph-mon[47734]: pgmap v3303: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:26.055 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:26.056 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:26.086 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:26.086 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:27 vm10.local ceph-mon[51861]: pgmap v3304: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:27 vm08.local ceph-mon[47734]: pgmap v3304: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:29 vm08.local ceph-mon[47734]: pgmap v3305: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:29 vm10.local ceph-mon[51861]: pgmap v3305: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:30 vm10.local ceph-mon[51861]: pgmap v3306: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:30 vm08.local ceph-mon[47734]: pgmap v3306: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:31.088 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:31.089 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:31.161 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:31.162 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:33 vm10.local ceph-mon[51861]: pgmap v3307: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:33 vm08.local ceph-mon[47734]: pgmap v3307: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:35 vm10.local ceph-mon[51861]: pgmap v3308: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:43:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:43:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:43:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:43:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:43:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:35 vm08.local ceph-mon[47734]: pgmap v3308: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:43:35.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:43:35.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:43:35.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:43:35.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:43:36.163 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:36.164 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:36.192 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:36.193 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:37 vm10.local ceph-mon[51861]: pgmap v3309: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:37.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:37 vm08.local ceph-mon[47734]: pgmap v3309: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:39 vm08.local ceph-mon[47734]: pgmap v3310: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:39.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:39 vm10.local ceph-mon[51861]: pgmap v3310: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:40 vm08.local ceph-mon[47734]: pgmap v3311: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:40 vm10.local ceph-mon[51861]: pgmap v3311: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:41.194 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:41.195 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:41.222 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:41.223 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:43.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:43 vm08.local ceph-mon[47734]: pgmap v3312: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:43 vm10.local ceph-mon[51861]: pgmap v3312: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:43:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:43:44.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:43:44.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:43:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:45 vm10.local ceph-mon[51861]: pgmap v3313: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:45 vm08.local ceph-mon[47734]: pgmap v3313: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:46.224 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:46.225 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:46.253 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:46.253 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:47 vm10.local ceph-mon[51861]: pgmap v3314: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:47 vm08.local ceph-mon[47734]: pgmap v3314: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:49 vm08.local ceph-mon[47734]: pgmap v3315: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:49 vm10.local ceph-mon[51861]: pgmap v3315: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:50.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:50 vm10.local ceph-mon[51861]: pgmap v3316: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:50 vm08.local ceph-mon[47734]: pgmap v3316: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:51.254 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:51.255 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:51.400 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:51.400 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:53 vm10.local ceph-mon[51861]: pgmap v3317: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:53 vm08.local ceph-mon[47734]: pgmap v3317: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:55 vm10.local ceph-mon[51861]: pgmap v3318: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:55 vm08.local ceph-mon[47734]: pgmap v3318: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:43:56.402 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:43:56.402 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:43:56.430 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:43:56.431 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:43:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:57 vm10.local ceph-mon[51861]: pgmap v3319: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:57 vm08.local ceph-mon[47734]: pgmap v3319: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:43:59 vm08.local ceph-mon[47734]: pgmap v3320: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:43:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:43:59 vm10.local ceph-mon[51861]: pgmap v3320: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:00.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:00 vm10.local ceph-mon[51861]: pgmap v3321: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:00 vm08.local ceph-mon[47734]: pgmap v3321: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:01.432 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:01.432 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:01.459 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:01.460 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:03 vm10.local ceph-mon[51861]: pgmap v3322: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:03 vm08.local ceph-mon[47734]: pgmap v3322: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:05 vm10.local ceph-mon[51861]: pgmap v3323: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:05 vm08.local ceph-mon[47734]: pgmap v3323: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:06.461 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:06.462 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:06.489 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:06.490 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:07 vm10.local ceph-mon[51861]: pgmap v3324: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:07 vm08.local ceph-mon[47734]: pgmap v3324: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:09 vm08.local ceph-mon[47734]: pgmap v3325: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:09 vm10.local ceph-mon[51861]: pgmap v3325: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:10.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:10 vm10.local ceph-mon[51861]: pgmap v3326: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:10 vm08.local ceph-mon[47734]: pgmap v3326: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:11.491 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:11.491 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:11.623 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:11.623 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:13 vm10.local ceph-mon[51861]: pgmap v3327: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:13.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:13 vm08.local ceph-mon[47734]: pgmap v3327: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:15 vm10.local ceph-mon[51861]: pgmap v3328: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:15 vm08.local ceph-mon[47734]: pgmap v3328: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:16.624 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:16.625 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:16.653 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:16.654 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:17 vm10.local ceph-mon[51861]: pgmap v3329: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:17 vm08.local ceph-mon[47734]: pgmap v3329: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:19 vm08.local ceph-mon[47734]: pgmap v3330: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:19.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:19 vm10.local ceph-mon[51861]: pgmap v3330: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:20.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:20 vm10.local ceph-mon[51861]: pgmap v3331: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:20 vm08.local ceph-mon[47734]: pgmap v3331: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:21.655 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:21.656 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:21.684 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:21.685 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:23 vm10.local ceph-mon[51861]: pgmap v3332: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:23 vm08.local ceph-mon[47734]: pgmap v3332: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:25 vm10.local ceph-mon[51861]: pgmap v3333: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:25 vm08.local ceph-mon[47734]: pgmap v3333: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:26.687 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:26.687 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:26.716 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:26.716 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:27 vm10.local ceph-mon[51861]: pgmap v3334: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:27 vm08.local ceph-mon[47734]: pgmap v3334: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:29 vm08.local ceph-mon[47734]: pgmap v3335: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:29 vm10.local ceph-mon[51861]: pgmap v3335: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:30.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:30 vm10.local ceph-mon[51861]: pgmap v3336: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:30 vm08.local ceph-mon[47734]: pgmap v3336: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:31.718 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:31.718 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:31.747 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:31.747 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:33 vm08.local ceph-mon[47734]: pgmap v3337: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:33.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:33 vm10.local ceph-mon[51861]: pgmap v3337: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:34.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:34 vm08.local ceph-mon[47734]: pgmap v3338: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:34.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:34 vm10.local ceph-mon[51861]: pgmap v3338: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:36.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:44:36.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:44:36.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:44:36.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:44:36.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:35 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:44:36.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:44:36.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:44:36.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:44:36.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:44:36.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:35 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:44:36.749 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:36.749 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:36.777 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:36.777 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:37.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:36 vm08.local ceph-mon[47734]: pgmap v3339: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:37.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:36 vm10.local ceph-mon[51861]: pgmap v3339: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:39 vm08.local ceph-mon[47734]: pgmap v3340: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:39 vm10.local ceph-mon[51861]: pgmap v3340: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:40 vm08.local ceph-mon[47734]: pgmap v3341: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:40 vm10.local ceph-mon[51861]: pgmap v3341: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:41.778 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:41.779 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:41.828 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:41.829 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:43 vm10.local ceph-mon[51861]: pgmap v3342: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:43 vm08.local ceph-mon[47734]: pgmap v3342: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:44:44.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:44 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:44:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:44:44.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:44 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:44:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:45 vm10.local ceph-mon[51861]: pgmap v3343: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:45 vm08.local ceph-mon[47734]: pgmap v3343: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:46.830 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:46.830 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:46.889 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:46.889 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:47 vm10.local ceph-mon[51861]: pgmap v3344: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:47 vm08.local ceph-mon[47734]: pgmap v3344: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:49 vm08.local ceph-mon[47734]: pgmap v3345: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:49.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:49 vm10.local ceph-mon[51861]: pgmap v3345: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:50 vm08.local ceph-mon[47734]: pgmap v3346: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:50 vm10.local ceph-mon[51861]: pgmap v3346: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:51.890 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:51.891 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:51.917 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:51.917 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:53 vm10.local ceph-mon[51861]: pgmap v3347: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:53 vm08.local ceph-mon[47734]: pgmap v3347: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:55 vm10.local ceph-mon[51861]: pgmap v3348: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:55 vm08.local ceph-mon[47734]: pgmap v3348: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:44:56.919 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:44:56.920 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:44:56.958 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:44:56.959 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:44:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:57 vm10.local ceph-mon[51861]: pgmap v3349: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:57 vm08.local ceph-mon[47734]: pgmap v3349: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:44:59 vm08.local ceph-mon[47734]: pgmap v3350: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:44:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:44:59 vm10.local ceph-mon[51861]: pgmap v3350: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:00 vm08.local ceph-mon[47734]: pgmap v3351: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:00 vm10.local ceph-mon[51861]: pgmap v3351: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:01.960 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:01.961 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:01.988 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:01.988 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:03 vm10.local ceph-mon[51861]: pgmap v3352: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:03 vm08.local ceph-mon[47734]: pgmap v3352: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:05 vm10.local ceph-mon[51861]: pgmap v3353: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:05.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:05 vm08.local ceph-mon[47734]: pgmap v3353: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:06.990 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:06.990 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:07.016 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:07.017 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:07 vm10.local ceph-mon[51861]: pgmap v3354: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:07 vm08.local ceph-mon[47734]: pgmap v3354: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:09 vm08.local ceph-mon[47734]: pgmap v3355: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:09 vm10.local ceph-mon[51861]: pgmap v3355: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:10 vm08.local ceph-mon[47734]: pgmap v3356: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:10 vm10.local ceph-mon[51861]: pgmap v3356: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:12.018 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:12.019 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:12.045 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:12.045 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:13 vm10.local ceph-mon[51861]: pgmap v3357: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:13.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:13 vm08.local ceph-mon[47734]: pgmap v3357: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:15 vm10.local ceph-mon[51861]: pgmap v3358: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:15.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:15 vm08.local ceph-mon[47734]: pgmap v3358: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:17.046 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:17.047 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:17.073 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:17.073 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:17 vm10.local ceph-mon[51861]: pgmap v3359: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:17.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:17 vm08.local ceph-mon[47734]: pgmap v3359: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:19 vm08.local ceph-mon[47734]: pgmap v3360: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:19.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:19 vm10.local ceph-mon[51861]: pgmap v3360: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:20 vm08.local ceph-mon[47734]: pgmap v3361: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:20 vm10.local ceph-mon[51861]: pgmap v3361: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:22.074 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:22.075 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:22.101 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:22.101 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:23 vm10.local ceph-mon[51861]: pgmap v3362: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:23 vm08.local ceph-mon[47734]: pgmap v3362: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:25 vm10.local ceph-mon[51861]: pgmap v3363: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:25.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:25 vm08.local ceph-mon[47734]: pgmap v3363: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:27.103 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:27.103 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:27.137 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:27.138 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:27 vm10.local ceph-mon[51861]: pgmap v3364: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:27 vm08.local ceph-mon[47734]: pgmap v3364: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:29 vm08.local ceph-mon[47734]: pgmap v3365: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:29 vm10.local ceph-mon[51861]: pgmap v3365: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:30 vm08.local ceph-mon[47734]: pgmap v3366: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:30 vm10.local ceph-mon[51861]: pgmap v3366: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:32.140 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:32.140 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:32.167 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:32.168 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:33 vm10.local ceph-mon[51861]: pgmap v3367: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:33 vm08.local ceph-mon[47734]: pgmap v3367: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:35.591 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:35 vm08.local ceph-mon[47734]: pgmap v3368: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:35.594 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:35 vm10.local ceph-mon[51861]: pgmap v3368: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:45:36.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:45:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:45:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:45:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:45:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:45:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:45:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:45:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:45:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:45:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:45:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:45:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:45:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:45:37.170 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:37.170 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:37.197 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:37.198 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:37 vm10.local ceph-mon[51861]: pgmap v3369: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:37.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:37 vm08.local ceph-mon[47734]: pgmap v3369: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:39 vm08.local ceph-mon[47734]: pgmap v3370: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:39 vm10.local ceph-mon[51861]: pgmap v3370: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:40 vm08.local ceph-mon[47734]: pgmap v3371: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:40 vm10.local ceph-mon[51861]: pgmap v3371: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:42.199 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:42.200 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:42.228 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:42.229 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:43 vm10.local ceph-mon[51861]: pgmap v3372: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:45:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:45:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:43 vm08.local ceph-mon[47734]: pgmap v3372: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:45:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:45:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:45 vm10.local ceph-mon[51861]: pgmap v3373: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:45 vm08.local ceph-mon[47734]: pgmap v3373: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:47.230 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:47.231 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:47.258 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:47.258 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:47 vm08.local ceph-mon[47734]: pgmap v3374: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:47.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:47 vm10.local ceph-mon[51861]: pgmap v3374: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:49.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:49 vm08.local ceph-mon[47734]: pgmap v3375: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:49 vm10.local ceph-mon[51861]: pgmap v3375: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:50 vm08.local ceph-mon[47734]: pgmap v3376: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:50 vm10.local ceph-mon[51861]: pgmap v3376: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:52.260 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:52.260 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:52.286 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:52.286 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:53 vm10.local ceph-mon[51861]: pgmap v3377: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:53 vm08.local ceph-mon[47734]: pgmap v3377: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:55 vm10.local ceph-mon[51861]: pgmap v3378: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:55.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:55 vm08.local ceph-mon[47734]: pgmap v3378: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:45:57.287 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:45:57.288 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:45:57.314 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:45:57.315 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:45:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:57 vm10.local ceph-mon[51861]: pgmap v3379: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:57 vm08.local ceph-mon[47734]: pgmap v3379: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:45:59 vm08.local ceph-mon[47734]: pgmap v3380: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:45:59.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:45:59 vm10.local ceph-mon[51861]: pgmap v3380: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:00 vm08.local ceph-mon[47734]: pgmap v3381: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:00 vm10.local ceph-mon[51861]: pgmap v3381: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:02.316 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:02.317 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:02.344 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:02.344 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:03.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:03 vm10.local ceph-mon[51861]: pgmap v3382: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:03 vm08.local ceph-mon[47734]: pgmap v3382: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:05 vm10.local ceph-mon[51861]: pgmap v3383: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:05.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:05 vm08.local ceph-mon[47734]: pgmap v3383: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:07.346 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:07.346 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:07.421 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:07.421 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:07 vm10.local ceph-mon[51861]: pgmap v3384: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:07 vm08.local ceph-mon[47734]: pgmap v3384: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:09 vm08.local ceph-mon[47734]: pgmap v3385: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:09 vm10.local ceph-mon[51861]: pgmap v3385: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:10 vm08.local ceph-mon[47734]: pgmap v3386: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:10 vm10.local ceph-mon[51861]: pgmap v3386: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:12.423 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:12.424 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:12.453 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:12.453 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:13 vm10.local ceph-mon[51861]: pgmap v3387: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:13.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:13 vm08.local ceph-mon[47734]: pgmap v3387: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:15 vm10.local ceph-mon[51861]: pgmap v3388: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:15.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:15 vm08.local ceph-mon[47734]: pgmap v3388: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:17.455 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:17.455 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:17.482 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:17.483 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:17 vm10.local ceph-mon[51861]: pgmap v3389: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:17 vm08.local ceph-mon[47734]: pgmap v3389: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:19 vm08.local ceph-mon[47734]: pgmap v3390: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:19 vm10.local ceph-mon[51861]: pgmap v3390: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:20 vm08.local ceph-mon[47734]: pgmap v3391: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:20 vm10.local ceph-mon[51861]: pgmap v3391: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:22.484 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:22.485 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:22.514 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:22.514 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:23 vm10.local ceph-mon[51861]: pgmap v3392: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:23 vm08.local ceph-mon[47734]: pgmap v3392: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:25 vm10.local ceph-mon[51861]: pgmap v3393: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:25.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:25 vm08.local ceph-mon[47734]: pgmap v3393: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:27.516 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:27.516 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:27.543 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:27.543 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:27 vm10.local ceph-mon[51861]: pgmap v3394: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:27 vm08.local ceph-mon[47734]: pgmap v3394: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:29 vm08.local ceph-mon[47734]: pgmap v3395: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:29 vm10.local ceph-mon[51861]: pgmap v3395: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:30 vm08.local ceph-mon[47734]: pgmap v3396: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:30 vm10.local ceph-mon[51861]: pgmap v3396: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:32.544 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:32.545 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:32.571 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:32.572 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:33 vm10.local ceph-mon[51861]: pgmap v3397: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:33 vm08.local ceph-mon[47734]: pgmap v3397: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:35.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:35 vm10.local ceph-mon[51861]: pgmap v3398: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:35.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:35 vm08.local ceph-mon[47734]: pgmap v3398: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:46:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:46:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:46:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:46:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:46:36.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:46:36.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:46:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:46:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:46:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm08", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:46:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:46:36.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:46:37.573 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:37.574 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:37.601 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:37.601 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:37 vm10.local ceph-mon[51861]: pgmap v3399: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:37.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:37 vm08.local ceph-mon[47734]: pgmap v3399: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:39 vm08.local ceph-mon[47734]: pgmap v3400: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:39.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:39 vm10.local ceph-mon[51861]: pgmap v3400: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:41.057 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:40 vm08.local ceph-mon[47734]: pgmap v3401: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:40 vm10.local ceph-mon[51861]: pgmap v3401: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:42.603 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:42.603 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:42.630 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:42.631 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:43 vm10.local ceph-mon[51861]: pgmap v3402: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:46:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:46:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:43 vm08.local ceph-mon[47734]: pgmap v3402: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:46:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:46:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:45 vm10.local ceph-mon[51861]: pgmap v3403: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:45 vm08.local ceph-mon[47734]: pgmap v3403: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:46.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:46 vm08.local ceph-mon[47734]: pgmap v3404: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:46.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:46 vm10.local ceph-mon[51861]: pgmap v3404: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:47.632 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:47.633 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:47.659 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:47.660 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:49 vm08.local ceph-mon[47734]: pgmap v3405: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:49.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:49 vm10.local ceph-mon[51861]: pgmap v3405: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:51.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:50 vm08.local ceph-mon[47734]: pgmap v3406: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:50 vm10.local ceph-mon[51861]: pgmap v3406: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:52.662 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:52.662 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:52.725 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:52.726 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:53 vm10.local ceph-mon[51861]: pgmap v3407: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:53 vm08.local ceph-mon[47734]: pgmap v3407: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:55.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:55 vm08.local ceph-mon[47734]: pgmap v3408: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:55 vm10.local ceph-mon[51861]: pgmap v3408: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:46:57.727 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:46:57.728 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:46:57.757 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:46:57.758 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:46:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:57 vm08.local ceph-mon[47734]: pgmap v3409: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:57.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:57 vm10.local ceph-mon[51861]: pgmap v3409: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:46:59 vm08.local ceph-mon[47734]: pgmap v3410: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:46:59.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:46:59 vm10.local ceph-mon[51861]: pgmap v3410: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:00 vm08.local ceph-mon[47734]: pgmap v3411: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:00 vm10.local ceph-mon[51861]: pgmap v3411: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:02.759 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:02.759 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:02.796 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:02.796 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:03 vm10.local ceph-mon[51861]: pgmap v3412: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:03 vm08.local ceph-mon[47734]: pgmap v3412: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:05 vm10.local ceph-mon[51861]: pgmap v3413: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:05.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:05 vm08.local ceph-mon[47734]: pgmap v3413: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:07 vm10.local ceph-mon[51861]: pgmap v3414: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:07.797 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:07.798 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:07 vm08.local ceph-mon[47734]: pgmap v3414: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:07.827 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:07.828 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:09 vm08.local ceph-mon[47734]: pgmap v3415: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:09.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:09 vm10.local ceph-mon[51861]: pgmap v3415: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:11.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:10 vm08.local ceph-mon[47734]: pgmap v3416: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:10 vm10.local ceph-mon[51861]: pgmap v3416: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:12.830 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:12.830 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:12.858 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:12.859 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:13 vm10.local ceph-mon[51861]: pgmap v3417: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:13.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:13 vm08.local ceph-mon[47734]: pgmap v3417: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:15 vm10.local ceph-mon[51861]: pgmap v3418: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:15 vm08.local ceph-mon[47734]: pgmap v3418: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:17 vm10.local ceph-mon[51861]: pgmap v3419: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:17.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:17 vm08.local ceph-mon[47734]: pgmap v3419: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:17.860 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:17.860 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:17.887 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:17.888 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:18.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:18 vm08.local ceph-mon[47734]: pgmap v3420: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:18.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:18 vm10.local ceph-mon[51861]: pgmap v3420: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:21.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:20 vm08.local ceph-mon[47734]: pgmap v3421: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:20 vm10.local ceph-mon[51861]: pgmap v3421: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:22.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:22 vm08.local ceph-mon[47734]: pgmap v3422: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:22.889 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:22.890 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:22.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:22 vm10.local ceph-mon[51861]: pgmap v3422: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:22.917 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:22.917 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:25 vm08.local ceph-mon[47734]: pgmap v3423: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:25.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:25 vm10.local ceph-mon[51861]: pgmap v3423: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:27 vm08.local ceph-mon[47734]: pgmap v3424: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:27.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:27 vm10.local ceph-mon[51861]: pgmap v3424: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:27.919 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:27.919 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:28.009 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:28.010 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:29.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:29 vm08.local ceph-mon[47734]: pgmap v3425: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:29.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:29 vm10.local ceph-mon[51861]: pgmap v3425: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:31.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:30 vm08.local ceph-mon[47734]: pgmap v3426: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:30 vm10.local ceph-mon[51861]: pgmap v3426: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:33.011 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:33.012 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:33.041 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:33.041 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:33.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:33 vm10.local ceph-mon[51861]: pgmap v3427: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:33 vm08.local ceph-mon[47734]: pgmap v3427: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:35 vm10.local ceph-mon[51861]: pgmap v3428: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:35 vm08.local ceph-mon[47734]: pgmap v3428: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:36.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:47:36.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:47:36.409 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:36 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:47:36.472 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:47:36.472 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:47:36.472 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:36 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:47:37.379 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:37 vm08.local ceph-mon[47734]: pgmap v3429: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:37.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:37 vm10.local ceph-mon[51861]: pgmap v3429: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:38.043 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:38.043 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:38.078 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:38.079 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:38.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:38 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:47:38.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:38 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:47:38.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:38 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:47:38.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:38 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:47:38.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:38 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:47:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:38 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:47:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:38 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T15:47:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:38 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:47:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:38 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:47:38.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:38 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:47:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:39 vm08.local ceph-mon[47734]: pgmap v3430: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:39 vm10.local ceph-mon[51861]: pgmap v3430: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:41.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:40 vm08.local ceph-mon[47734]: pgmap v3431: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:40 vm10.local ceph-mon[51861]: pgmap v3431: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:43.080 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:43.081 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:43.107 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:43.108 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:43 vm10.local ceph-mon[51861]: pgmap v3432: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:47:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:47:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:43 vm08.local ceph-mon[47734]: pgmap v3432: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:47:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:47:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:45 vm08.local ceph-mon[47734]: pgmap v3433: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:45.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:45 vm10.local ceph-mon[51861]: pgmap v3433: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:47 vm08.local ceph-mon[47734]: pgmap v3434: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:47.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:47 vm10.local ceph-mon[51861]: pgmap v3434: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:48.109 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:48.110 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:48.138 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:48.138 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:49 vm08.local ceph-mon[47734]: pgmap v3435: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:49.769 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:49 vm10.local ceph-mon[51861]: pgmap v3435: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:50 vm10.local ceph-mon[51861]: pgmap v3436: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:50 vm08.local ceph-mon[47734]: pgmap v3436: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:53.140 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:53.140 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:53.253 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:53.253 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:53 vm10.local ceph-mon[51861]: pgmap v3437: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:53 vm08.local ceph-mon[47734]: pgmap v3437: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:55 vm08.local ceph-mon[47734]: pgmap v3438: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:55 vm10.local ceph-mon[51861]: pgmap v3438: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:47:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:57 vm08.local ceph-mon[47734]: pgmap v3439: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:57.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:57 vm10.local ceph-mon[51861]: pgmap v3439: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:58.256 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:47:58.257 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:47:58.388 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:47:58.389 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:47:58.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:47:58 vm08.local ceph-mon[47734]: pgmap v3440: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:47:58.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:47:58 vm10.local ceph-mon[51861]: pgmap v3440: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:01.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:00 vm08.local ceph-mon[47734]: pgmap v3441: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:00 vm10.local ceph-mon[51861]: pgmap v3441: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:03.390 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:03.391 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:03.419 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:03.419 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:03 vm10.local ceph-mon[51861]: pgmap v3442: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:03 vm08.local ceph-mon[47734]: pgmap v3442: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:05 vm10.local ceph-mon[51861]: pgmap v3443: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:05 vm08.local ceph-mon[47734]: pgmap v3443: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:07 vm10.local ceph-mon[51861]: pgmap v3444: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:07 vm08.local ceph-mon[47734]: pgmap v3444: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:08.421 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:08.422 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:08.452 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:08.452 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:09 vm08.local ceph-mon[47734]: pgmap v3445: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:09.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:09 vm10.local ceph-mon[51861]: pgmap v3445: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:11.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:10 vm10.local ceph-mon[51861]: pgmap v3446: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:10 vm08.local ceph-mon[47734]: pgmap v3446: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:13.453 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:13.454 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:13.480 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:13.481 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:13 vm10.local ceph-mon[51861]: pgmap v3447: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:13.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:13 vm08.local ceph-mon[47734]: pgmap v3447: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:15 vm10.local ceph-mon[51861]: pgmap v3448: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:15 vm08.local ceph-mon[47734]: pgmap v3448: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:17.479 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:17 vm08.local ceph-mon[47734]: pgmap v3449: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:17 vm10.local ceph-mon[51861]: pgmap v3449: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:18.482 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:18.483 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:18.513 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:18.514 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:19 vm08.local ceph-mon[47734]: pgmap v3450: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:19 vm10.local ceph-mon[51861]: pgmap v3450: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:21.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:20 vm10.local ceph-mon[51861]: pgmap v3451: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:20 vm08.local ceph-mon[47734]: pgmap v3451: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:23.515 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:23.516 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:23.543 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:23.543 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:23 vm10.local ceph-mon[51861]: pgmap v3452: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:23 vm08.local ceph-mon[47734]: pgmap v3452: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:25 vm10.local ceph-mon[51861]: pgmap v3453: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:25 vm08.local ceph-mon[47734]: pgmap v3453: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:27 vm10.local ceph-mon[51861]: pgmap v3454: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:27 vm08.local ceph-mon[47734]: pgmap v3454: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:28.545 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:28.545 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:28.573 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:28.573 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:29 vm08.local ceph-mon[47734]: pgmap v3455: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:29.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:29 vm10.local ceph-mon[51861]: pgmap v3455: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:31.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:30 vm10.local ceph-mon[51861]: pgmap v3456: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:30 vm08.local ceph-mon[47734]: pgmap v3456: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:33.575 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:33.575 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:33.603 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:33.604 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:33 vm10.local ceph-mon[51861]: pgmap v3457: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:33 vm08.local ceph-mon[47734]: pgmap v3457: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:35 vm10.local ceph-mon[51861]: pgmap v3458: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:35 vm08.local ceph-mon[47734]: pgmap v3458: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:37 vm10.local ceph-mon[51861]: pgmap v3459: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:37.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:37 vm08.local ceph-mon[47734]: pgmap v3459: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:38.606 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:38.606 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:38.640 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:38.641 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:38 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:48:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:38 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:48:38.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:38 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:48:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:38 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:48:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:38 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:48:38.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:38 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:48:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:39 vm08.local ceph-mon[47734]: pgmap v3460: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:48:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:48:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:39 vm10.local ceph-mon[51861]: pgmap v3460: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:48:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:48:41.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:40 vm10.local ceph-mon[51861]: pgmap v3461: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:40 vm08.local ceph-mon[47734]: pgmap v3461: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:43.642 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:43.642 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:43 vm10.local ceph-mon[51861]: pgmap v3462: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:48:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:48:43.672 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:43.673 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:43 vm08.local ceph-mon[47734]: pgmap v3462: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:48:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:48:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:45 vm10.local ceph-mon[51861]: pgmap v3463: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:45.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:45 vm08.local ceph-mon[47734]: pgmap v3463: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:47 vm10.local ceph-mon[51861]: pgmap v3464: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:47.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:47 vm08.local ceph-mon[47734]: pgmap v3464: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:48.674 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:48.675 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:48.701 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:48.701 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:49 vm08.local ceph-mon[47734]: pgmap v3465: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:49.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:49 vm10.local ceph-mon[51861]: pgmap v3465: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:51.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:50 vm10.local ceph-mon[51861]: pgmap v3466: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:50 vm08.local ceph-mon[47734]: pgmap v3466: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:48:53.703 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:53.703 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:53.731 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:53.731 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:53 vm08.local ceph-mon[47734]: pgmap v3467: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:53.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:53 vm10.local ceph-mon[51861]: pgmap v3467: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:55 vm08.local ceph-mon[47734]: pgmap v3468: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:55.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:55 vm10.local ceph-mon[51861]: pgmap v3468: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:48:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:57 vm08.local ceph-mon[47734]: pgmap v3469: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:48:57.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:57 vm10.local ceph-mon[51861]: pgmap v3469: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:48:58.733 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:48:58.733 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:48:58.763 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:48:58.764 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:48:59.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:48:59 vm08.local ceph-mon[47734]: pgmap v3470: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:48:59.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:48:59 vm10.local ceph-mon[51861]: pgmap v3470: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:49:01.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:00 vm10.local ceph-mon[51861]: pgmap v3471: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:00 vm08.local ceph-mon[47734]: pgmap v3471: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:03 vm10.local ceph-mon[51861]: pgmap v3472: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:49:03.765 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:03.766 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:03.795 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:03.795 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:03 vm08.local ceph-mon[47734]: pgmap v3472: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T15:49:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:05 vm08.local ceph-mon[47734]: pgmap v3473: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:05.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:05 vm10.local ceph-mon[51861]: pgmap v3473: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:07.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:07 vm08.local ceph-mon[47734]: pgmap v3474: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:07.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:07 vm10.local ceph-mon[51861]: pgmap v3474: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:08.797 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:08.797 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:08.823 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:08.824 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:09.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:09 vm08.local ceph-mon[47734]: pgmap v3475: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:09.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:09 vm10.local ceph-mon[51861]: pgmap v3475: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:10 vm08.local ceph-mon[47734]: pgmap v3476: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:10 vm10.local ceph-mon[51861]: pgmap v3476: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:13 vm10.local ceph-mon[51861]: pgmap v3477: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:13.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:13 vm08.local ceph-mon[47734]: pgmap v3477: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:13.825 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:13.826 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:13.852 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:13.852 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:15 vm10.local ceph-mon[51861]: pgmap v3478: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:15 vm08.local ceph-mon[47734]: pgmap v3478: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:17 vm10.local ceph-mon[51861]: pgmap v3479: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:17.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:17 vm08.local ceph-mon[47734]: pgmap v3479: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:18.854 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:18.855 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:18.883 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:18.884 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:19 vm08.local ceph-mon[47734]: pgmap v3480: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:19 vm10.local ceph-mon[51861]: pgmap v3480: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:20 vm08.local ceph-mon[47734]: pgmap v3481: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:20 vm10.local ceph-mon[51861]: pgmap v3481: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:23 vm10.local ceph-mon[51861]: pgmap v3482: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:23 vm08.local ceph-mon[47734]: pgmap v3482: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:23.885 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:23.886 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:23.913 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:23.914 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:25.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:25 vm10.local ceph-mon[51861]: pgmap v3483: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:25.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:25 vm08.local ceph-mon[47734]: pgmap v3483: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:27 vm10.local ceph-mon[51861]: pgmap v3484: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:27.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:27 vm08.local ceph-mon[47734]: pgmap v3484: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:28.915 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:28.915 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:28.943 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:28.943 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:29.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:29 vm08.local ceph-mon[47734]: pgmap v3485: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:29.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:29 vm10.local ceph-mon[51861]: pgmap v3485: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:31.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:30 vm08.local ceph-mon[47734]: pgmap v3486: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:30 vm10.local ceph-mon[51861]: pgmap v3486: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:33.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:33 vm10.local ceph-mon[51861]: pgmap v3487: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:33 vm08.local ceph-mon[47734]: pgmap v3487: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:33.945 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:33.945 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:33.971 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:33.972 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:35.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:35 vm10.local ceph-mon[51861]: pgmap v3488: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:35.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:35 vm08.local ceph-mon[47734]: pgmap v3488: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:37.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:37 vm10.local ceph-mon[51861]: pgmap v3489: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:37.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:37 vm08.local ceph-mon[47734]: pgmap v3489: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:38.973 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:38.974 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:39.005 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:39.006 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:39.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:39 vm08.local ceph-mon[47734]: pgmap v3490: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:49:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:49:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:49:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:49:39.557 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:49:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:39 vm10.local ceph-mon[51861]: pgmap v3490: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:39.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:49:39.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:49:39.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:49:39.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:49:39.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:49:41.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:40 vm08.local ceph-mon[47734]: pgmap v3491: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:40 vm10.local ceph-mon[51861]: pgmap v3491: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:43 vm10.local ceph-mon[51861]: pgmap v3492: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:49:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:49:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:43 vm08.local ceph-mon[47734]: pgmap v3492: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:49:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:49:44.008 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:44.008 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:44.037 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:44.038 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:45 vm10.local ceph-mon[51861]: pgmap v3493: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:45 vm08.local ceph-mon[47734]: pgmap v3493: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:47 vm10.local ceph-mon[51861]: pgmap v3494: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:47.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:47 vm08.local ceph-mon[47734]: pgmap v3494: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:49.039 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:49.040 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:49.070 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:49.071 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:49.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:49 vm08.local ceph-mon[47734]: pgmap v3495: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:49 vm10.local ceph-mon[51861]: pgmap v3495: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:50 vm08.local ceph-mon[47734]: pgmap v3496: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:50 vm10.local ceph-mon[51861]: pgmap v3496: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:53 vm10.local ceph-mon[51861]: pgmap v3497: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:53.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:53 vm08.local ceph-mon[47734]: pgmap v3497: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:54.072 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:54.073 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:54.100 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:54.100 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:55 vm10.local ceph-mon[51861]: pgmap v3498: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:55 vm08.local ceph-mon[47734]: pgmap v3498: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:49:57.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:57 vm10.local ceph-mon[51861]: pgmap v3499: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:57.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:57 vm08.local ceph-mon[47734]: pgmap v3499: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:59.102 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:49:59.103 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:49:59.129 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:49:59.130 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:49:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:49:59 vm08.local ceph-mon[47734]: pgmap v3500: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:49:59.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:49:59 vm10.local ceph-mon[51861]: pgmap v3500: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:00 vm08.local ceph-mon[47734]: overall HEALTH_OK 2026-03-09T15:50:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:00 vm08.local ceph-mon[47734]: pgmap v3501: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:00 vm10.local ceph-mon[51861]: overall HEALTH_OK 2026-03-09T15:50:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:00 vm10.local ceph-mon[51861]: pgmap v3501: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:03.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:03 vm10.local ceph-mon[51861]: pgmap v3502: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:03.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:03 vm08.local ceph-mon[47734]: pgmap v3502: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:04.131 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:04.132 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:04.160 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:04.160 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:05 vm10.local ceph-mon[51861]: pgmap v3503: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:05.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:05 vm08.local ceph-mon[47734]: pgmap v3503: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:07.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:07 vm10.local ceph-mon[51861]: pgmap v3504: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:07.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:07 vm08.local ceph-mon[47734]: pgmap v3504: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:09.161 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:09.162 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:09.190 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:09.191 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:09.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:09 vm08.local ceph-mon[47734]: pgmap v3505: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:09.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:09 vm10.local ceph-mon[51861]: pgmap v3505: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:11.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:10 vm08.local ceph-mon[47734]: pgmap v3506: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:11.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:10 vm10.local ceph-mon[51861]: pgmap v3506: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:13.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:13 vm10.local ceph-mon[51861]: pgmap v3507: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:13.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:13 vm08.local ceph-mon[47734]: pgmap v3507: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:14.193 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:14.193 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:14.220 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:14.221 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:15.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:15 vm10.local ceph-mon[51861]: pgmap v3508: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:15.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:15 vm08.local ceph-mon[47734]: pgmap v3508: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:17.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:17 vm10.local ceph-mon[51861]: pgmap v3509: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:17.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:17 vm08.local ceph-mon[47734]: pgmap v3509: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:19.222 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:19.222 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:19.248 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:19.249 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:19.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:19 vm08.local ceph-mon[47734]: pgmap v3510: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:19.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:19 vm10.local ceph-mon[51861]: pgmap v3510: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:21.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:21 vm08.local ceph-mon[47734]: pgmap v3511: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:21.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:21 vm10.local ceph-mon[51861]: pgmap v3511: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:23.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:23 vm10.local ceph-mon[51861]: pgmap v3512: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:23.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:23 vm08.local ceph-mon[47734]: pgmap v3512: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:24.250 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:24.251 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:24.546 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:24.546 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:25.056 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:24 vm08.local ceph-mon[47734]: pgmap v3513: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:25.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:24 vm10.local ceph-mon[51861]: pgmap v3513: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:27.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:27 vm10.local ceph-mon[51861]: pgmap v3514: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:27.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:27 vm08.local ceph-mon[47734]: pgmap v3514: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:29.548 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:29.548 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:29.713 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:29.714 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:29.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:29 vm08.local ceph-mon[47734]: pgmap v3515: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:29.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:29 vm10.local ceph-mon[51861]: pgmap v3515: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:31.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:31 vm08.local ceph-mon[47734]: pgmap v3516: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:31.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:31 vm10.local ceph-mon[51861]: pgmap v3516: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:33.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:33 vm08.local ceph-mon[47734]: pgmap v3517: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:33.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:33 vm10.local ceph-mon[51861]: pgmap v3517: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:34.715 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:34.716 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:34.743 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:34.743 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:35.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:35 vm08.local ceph-mon[47734]: pgmap v3518: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:35.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:35 vm10.local ceph-mon[51861]: pgmap v3518: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:37.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:37 vm08.local ceph-mon[47734]: pgmap v3519: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:37.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:37 vm10.local ceph-mon[51861]: pgmap v3519: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:39.499 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:39 vm08.local ceph-mon[47734]: pgmap v3520: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:39.499 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:50:39.499 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:50:39.744 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:39.745 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:39.772 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:39.773 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:39.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:50:39.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:50:39.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:39 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:50:39.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:39 vm10.local ceph-mon[51861]: pgmap v3520: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:39.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T15:50:39.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T15:50:39.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T15:50:39.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:50:39.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:39 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' 2026-03-09T15:50:41.307 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:41 vm08.local ceph-mon[47734]: pgmap v3521: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:41.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:41 vm10.local ceph-mon[51861]: pgmap v3521: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:43 vm10.local ceph-mon[51861]: pgmap v3522: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:43.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:50:43.659 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:43 vm10.local ceph-mon[51861]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:50:43.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:43 vm08.local ceph-mon[47734]: pgmap v3522: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/trash_purge_schedule"}]: dispatch 2026-03-09T15:50:43.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:43 vm08.local ceph-mon[47734]: from='mgr.14232 192.168.123.108:0/4174987312' entity='mgr.vm08.siqces' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm08.siqces/mirror_snapshot_schedule"}]: dispatch 2026-03-09T15:50:44.774 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:44.774 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:44.800 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:44.801 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:45.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:45 vm10.local ceph-mon[51861]: pgmap v3523: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:45.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:45 vm08.local ceph-mon[47734]: pgmap v3523: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:47.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:47 vm10.local ceph-mon[51861]: pgmap v3524: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:47.807 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:47 vm08.local ceph-mon[47734]: pgmap v3524: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:49.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:49 vm08.local ceph-mon[47734]: pgmap v3525: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:49.802 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:49.802 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:49.864 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:49.864 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:49.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:49 vm10.local ceph-mon[51861]: pgmap v3525: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:51.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:51 vm08.local ceph-mon[47734]: pgmap v3526: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:51.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:51 vm10.local ceph-mon[51861]: pgmap v3526: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:53.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:53 vm10.local ceph-mon[51861]: pgmap v3527: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:53.808 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:53 vm08.local ceph-mon[47734]: pgmap v3527: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:54.866 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:54.866 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:54.892 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:54.893 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:50:55.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:55 vm10.local ceph-mon[51861]: pgmap v3528: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:55.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:55 vm08.local ceph-mon[47734]: pgmap v3528: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:50:57.806 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:57 vm08.local ceph-mon[47734]: pgmap v3529: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:57.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:57 vm10.local ceph-mon[51861]: pgmap v3529: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:59.556 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:50:59 vm08.local ceph-mon[47734]: pgmap v3530: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:59.894 INFO:teuthology.orchestra.run.vm08.stderr:++ hostname 2026-03-09T15:50:59.895 INFO:teuthology.orchestra.run.vm08.stderr:+ mount -t nfs vm08.local:/fake /mnt/foo -o sync 2026-03-09T15:50:59.908 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:50:59 vm10.local ceph-mon[51861]: pgmap v3530: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T15:50:59.921 INFO:teuthology.orchestra.run.vm08.stderr:mount.nfs: mount system call failed 2026-03-09T15:50:59.922 INFO:teuthology.orchestra.run.vm08.stderr:+ sleep 5 2026-03-09T15:51:01.306 INFO:journalctl@ceph.mon.vm08.vm08.stdout:Mar 09 15:51:01 vm08.local ceph-mon[47734]: pgmap v3531: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:51:01.408 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 15:51:01 vm10.local ceph-mon[51861]: pgmap v3531: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T15:51:02.068 DEBUG:teuthology.exit:Got signal 15; running 1 handler... 2026-03-09T15:51:02.076 DEBUG:teuthology.exit:Finished running handlers