2026-03-09T17:22:44.903 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-09T17:22:44.908 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T17:22:44.928 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/586 branch: squid description: orch/cephadm/mgr-nfs-upgrade/{0-centos_9.stream 1-bootstrap/17.2.0 1-start 2-nfs 3-upgrade-with-workload 4-final} email: null first_in_suite: false flavor: default job_id: '586' last_in_suite: false machine_type: vps name: kyr-2026-03-09_11:23:05-orch-squid-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_REFRESH_FAILED log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - osd.0 - osd.1 - osd.2 - osd.3 - client.0 - - host.b - osd.4 - osd.5 - osd.6 - osd.7 seed: 3443 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 targets: vm10.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBFasX5HBt4oY13DN/fjueIqjl1CPLLd1yhyfN7C3wmae6R/jlpxW6owZGGUnFRVM7Z2uo02bUO83vSBEAMvmDeY= vm11.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMn8sncN4GMjUWWgGWiF4EVrg7sA7PxXFuv3ilm8l7cPbenpH++Zep89cKR8wCW6BdY2xV/DueZUGPGOEU6tXQo= tasks: - cephadm: cephadm_branch: v17.2.0 cephadm_git_url: https://github.com/ceph/ceph image: quay.io/ceph/ceph:v17.2.0 roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph fs volume create foofs - cephadm.wait_for_service: service: mds.foofs - cephadm.shell: host.a: - ceph nfs cluster create foo --placement=2 || ceph nfs cluster create cephfs foo --placement=2 - ceph nfs export create cephfs --fsname foofs --clusterid foo --binding /fake || ceph nfs export create cephfs --fsname foofs --cluster-id foo --pseudo-path /fake - while ! ceph orch ls | grep nfs | grep 2/2 ; do sleep 1 ; done - vip.exec: host.a: - mkdir /mnt/foo - while ! mount -t nfs $(hostname):/fake /mnt/foo -o sync ; do sleep 5 ; done - echo test > /mnt/foo/testfile - sync - parallel: - upgrade-tasks - workload-tasks - vip.exec: host.a: - umount /mnt/foo - cephadm.shell: host.a: - ceph nfs cluster ls | grep foo - ceph nfs export ls foo --detailed - rados -p .nfs --all ls - - cephadm.shell: host.a: - 'set -ex [[ `ceph config get mgr mgr/cephadm/migration_current` -gt 2 ]] ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-09_11:23:05 tube: vps upgrade-tasks: sequential: - cephadm.shell: env: - sha1 host.a: - ceph config set mon mon_warn_on_insecure_global_id_reclaim false --force - ceph config set mon mon_warn_on_insecure_global_id_reclaim_allowed false --force - ceph config set global log_to_journald false --force - ceph mgr module enable nfs --force - ceph orch upgrade start --image quay.ceph.io/ceph-ci/ceph:$sha1 - cephadm.shell: env: - sha1 host.a: - while ceph orch upgrade status | jq '.in_progress' | grep true && ! ceph orch upgrade status | jq '.message' | grep Error ; do ceph orch ps ; ceph versions ; ceph orch upgrade status ; ceph health detail ; sleep 30 ; done - ceph orch ps - ceph versions - echo "wait for servicemap items w/ changing names to refresh" - sleep 60 - ceph orch ps - ceph orch upgrade status - ceph health detail - ceph versions - ceph versions | jq -e '.overall | length == 1' - ceph versions | jq -e '.overall | keys' | grep $sha1 - cephadm.wait_for_service: service: nfs.foo user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 workload-tasks: sequential: - exec: host.a: - cd /mnt/foo && dbench 5 -t 600 || true - umount /mnt/foo - while ! mount -t nfs $(hostname):/fake /mnt/foo ; do sleep 5 ; done - cd /mnt/foo && dbench 5 -t 5 2026-03-09T17:22:44.928 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa; will attempt to use it 2026-03-09T17:22:44.929 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa/tasks 2026-03-09T17:22:44.929 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-09T17:22:44.929 INFO:teuthology.task.internal:Checking packages... 2026-03-09T17:22:44.929 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-09T17:22:44.929 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-09T17:22:44.929 INFO:teuthology.packaging:ref: None 2026-03-09T17:22:44.929 INFO:teuthology.packaging:tag: None 2026-03-09T17:22:44.929 INFO:teuthology.packaging:branch: squid 2026-03-09T17:22:44.929 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T17:22:44.929 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-09T17:22:45.632 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-09T17:22:45.633 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-09T17:22:45.634 INFO:teuthology.task.internal:no buildpackages task found 2026-03-09T17:22:45.634 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-09T17:22:45.634 INFO:teuthology.task.internal:Saving configuration 2026-03-09T17:22:45.640 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-09T17:22:45.641 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-09T17:22:45.648 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm10.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/586', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 17:20:25.157557', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:0a', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBFasX5HBt4oY13DN/fjueIqjl1CPLLd1yhyfN7C3wmae6R/jlpxW6owZGGUnFRVM7Z2uo02bUO83vSBEAMvmDeY='} 2026-03-09T17:22:45.653 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm11.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/586', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 17:20:25.158337', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:0b', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMn8sncN4GMjUWWgGWiF4EVrg7sA7PxXFuv3ilm8l7cPbenpH++Zep89cKR8wCW6BdY2xV/DueZUGPGOEU6tXQo='} 2026-03-09T17:22:45.654 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-09T17:22:45.654 INFO:teuthology.task.internal:roles: ubuntu@vm10.local - ['host.a', 'osd.0', 'osd.1', 'osd.2', 'osd.3', 'client.0'] 2026-03-09T17:22:45.654 INFO:teuthology.task.internal:roles: ubuntu@vm11.local - ['host.b', 'osd.4', 'osd.5', 'osd.6', 'osd.7'] 2026-03-09T17:22:45.654 INFO:teuthology.run_tasks:Running task console_log... 2026-03-09T17:22:45.665 DEBUG:teuthology.task.console_log:vm10 does not support IPMI; excluding 2026-03-09T17:22:45.671 DEBUG:teuthology.task.console_log:vm11 does not support IPMI; excluding 2026-03-09T17:22:45.672 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7fb10b112170>, signals=[15]) 2026-03-09T17:22:45.672 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-09T17:22:45.672 INFO:teuthology.task.internal:Opening connections... 2026-03-09T17:22:45.673 DEBUG:teuthology.task.internal:connecting to ubuntu@vm10.local 2026-03-09T17:22:45.673 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm10.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:22:45.731 DEBUG:teuthology.task.internal:connecting to ubuntu@vm11.local 2026-03-09T17:22:45.732 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm11.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:22:45.791 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-09T17:22:45.793 DEBUG:teuthology.orchestra.run.vm10:> uname -m 2026-03-09T17:22:45.853 INFO:teuthology.orchestra.run.vm10.stdout:x86_64 2026-03-09T17:22:45.853 DEBUG:teuthology.orchestra.run.vm10:> cat /etc/os-release 2026-03-09T17:22:45.910 INFO:teuthology.orchestra.run.vm10.stdout:NAME="CentOS Stream" 2026-03-09T17:22:45.910 INFO:teuthology.orchestra.run.vm10.stdout:VERSION="9" 2026-03-09T17:22:45.910 INFO:teuthology.orchestra.run.vm10.stdout:ID="centos" 2026-03-09T17:22:45.910 INFO:teuthology.orchestra.run.vm10.stdout:ID_LIKE="rhel fedora" 2026-03-09T17:22:45.910 INFO:teuthology.orchestra.run.vm10.stdout:VERSION_ID="9" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:PLATFORM_ID="platform:el9" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:ANSI_COLOR="0;31" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:LOGO="fedora-logo-icon" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:HOME_URL="https://centos.org/" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T17:22:45.911 INFO:teuthology.orchestra.run.vm10.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T17:22:45.911 INFO:teuthology.lock.ops:Updating vm10.local on lock server 2026-03-09T17:22:45.915 DEBUG:teuthology.orchestra.run.vm11:> uname -m 2026-03-09T17:22:45.932 INFO:teuthology.orchestra.run.vm11.stdout:x86_64 2026-03-09T17:22:45.932 DEBUG:teuthology.orchestra.run.vm11:> cat /etc/os-release 2026-03-09T17:22:45.988 INFO:teuthology.orchestra.run.vm11.stdout:NAME="CentOS Stream" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:VERSION="9" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:ID="centos" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:ID_LIKE="rhel fedora" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:VERSION_ID="9" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:PLATFORM_ID="platform:el9" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:ANSI_COLOR="0;31" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:LOGO="fedora-logo-icon" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:HOME_URL="https://centos.org/" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T17:22:45.989 INFO:teuthology.orchestra.run.vm11.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T17:22:45.989 INFO:teuthology.lock.ops:Updating vm11.local on lock server 2026-03-09T17:22:45.994 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-09T17:22:45.996 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-09T17:22:45.997 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-09T17:22:45.997 DEBUG:teuthology.orchestra.run.vm10:> test '!' -e /home/ubuntu/cephtest 2026-03-09T17:22:45.999 DEBUG:teuthology.orchestra.run.vm11:> test '!' -e /home/ubuntu/cephtest 2026-03-09T17:22:46.044 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-09T17:22:46.046 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-09T17:22:46.046 DEBUG:teuthology.orchestra.run.vm10:> test -z $(ls -A /var/lib/ceph) 2026-03-09T17:22:46.053 DEBUG:teuthology.orchestra.run.vm11:> test -z $(ls -A /var/lib/ceph) 2026-03-09T17:22:46.069 INFO:teuthology.orchestra.run.vm10.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T17:22:46.103 INFO:teuthology.orchestra.run.vm11.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T17:22:46.104 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-09T17:22:46.116 DEBUG:teuthology.orchestra.run.vm10:> test -e /ceph-qa-ready 2026-03-09T17:22:46.134 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:22:46.326 DEBUG:teuthology.orchestra.run.vm11:> test -e /ceph-qa-ready 2026-03-09T17:22:46.340 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:22:46.546 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-09T17:22:46.547 INFO:teuthology.task.internal:Creating test directory... 2026-03-09T17:22:46.547 DEBUG:teuthology.orchestra.run.vm10:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T17:22:46.550 DEBUG:teuthology.orchestra.run.vm11:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T17:22:46.567 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-09T17:22:46.569 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-09T17:22:46.570 INFO:teuthology.task.internal:Creating archive directory... 2026-03-09T17:22:46.570 DEBUG:teuthology.orchestra.run.vm10:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T17:22:46.609 DEBUG:teuthology.orchestra.run.vm11:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T17:22:46.630 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-09T17:22:46.631 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-09T17:22:46.631 DEBUG:teuthology.orchestra.run.vm10:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T17:22:46.687 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:22:46.688 DEBUG:teuthology.orchestra.run.vm11:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T17:22:46.704 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:22:46.704 DEBUG:teuthology.orchestra.run.vm10:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T17:22:46.731 DEBUG:teuthology.orchestra.run.vm11:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T17:22:46.756 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:22:46.767 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:22:46.768 INFO:teuthology.orchestra.run.vm11.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:22:46.777 INFO:teuthology.orchestra.run.vm11.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T17:22:46.778 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-09T17:22:46.780 INFO:teuthology.task.internal:Configuring sudo... 2026-03-09T17:22:46.780 DEBUG:teuthology.orchestra.run.vm10:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T17:22:46.812 DEBUG:teuthology.orchestra.run.vm11:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T17:22:46.840 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-09T17:22:46.843 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-09T17:22:46.843 DEBUG:teuthology.orchestra.run.vm10:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T17:22:46.879 DEBUG:teuthology.orchestra.run.vm11:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T17:22:46.895 DEBUG:teuthology.orchestra.run.vm10:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T17:22:46.960 DEBUG:teuthology.orchestra.run.vm10:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T17:22:47.018 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:22:47.018 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T17:22:47.077 DEBUG:teuthology.orchestra.run.vm11:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T17:22:47.102 DEBUG:teuthology.orchestra.run.vm11:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T17:22:47.160 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T17:22:47.160 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T17:22:47.223 DEBUG:teuthology.orchestra.run.vm10:> sudo service rsyslog restart 2026-03-09T17:22:47.225 DEBUG:teuthology.orchestra.run.vm11:> sudo service rsyslog restart 2026-03-09T17:22:47.252 INFO:teuthology.orchestra.run.vm10.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T17:22:47.294 INFO:teuthology.orchestra.run.vm11.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T17:22:47.629 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-09T17:22:47.631 INFO:teuthology.task.internal:Starting timer... 2026-03-09T17:22:47.631 INFO:teuthology.run_tasks:Running task pcp... 2026-03-09T17:22:47.633 INFO:teuthology.run_tasks:Running task selinux... 2026-03-09T17:22:47.636 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-09T17:22:47.636 INFO:teuthology.task.selinux:Excluding vm10: VMs are not yet supported 2026-03-09T17:22:47.636 INFO:teuthology.task.selinux:Excluding vm11: VMs are not yet supported 2026-03-09T17:22:47.636 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-09T17:22:47.636 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-09T17:22:47.636 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-09T17:22:47.636 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-09T17:22:47.638 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-09T17:22:47.638 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-09T17:22:47.640 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-09T17:22:48.386 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-09T17:22:48.391 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-09T17:22:48.392 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryhu_qd_vb --limit vm10.local,vm11.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-09T17:25:45.146 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm10.local'), Remote(name='ubuntu@vm11.local')] 2026-03-09T17:25:45.147 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm10.local' 2026-03-09T17:25:45.147 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm10.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:25:45.211 DEBUG:teuthology.orchestra.run.vm10:> true 2026-03-09T17:25:45.293 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm10.local' 2026-03-09T17:25:45.293 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm11.local' 2026-03-09T17:25:45.294 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm11.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T17:25:45.357 DEBUG:teuthology.orchestra.run.vm11:> true 2026-03-09T17:25:45.436 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm11.local' 2026-03-09T17:25:45.436 INFO:teuthology.run_tasks:Running task clock... 2026-03-09T17:25:45.439 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-09T17:25:45.439 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T17:25:45.439 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T17:25:45.440 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T17:25:45.440 DEBUG:teuthology.orchestra.run.vm11:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T17:25:45.476 INFO:teuthology.orchestra.run.vm10.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T17:25:45.500 INFO:teuthology.orchestra.run.vm10.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T17:25:45.508 INFO:teuthology.orchestra.run.vm11.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T17:25:45.519 INFO:teuthology.orchestra.run.vm11.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T17:25:45.527 INFO:teuthology.orchestra.run.vm10.stderr:sudo: ntpd: command not found 2026-03-09T17:25:45.542 INFO:teuthology.orchestra.run.vm10.stdout:506 Cannot talk to daemon 2026-03-09T17:25:45.545 INFO:teuthology.orchestra.run.vm11.stderr:sudo: ntpd: command not found 2026-03-09T17:25:45.554 INFO:teuthology.orchestra.run.vm11.stdout:506 Cannot talk to daemon 2026-03-09T17:25:45.557 INFO:teuthology.orchestra.run.vm10.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T17:25:45.568 INFO:teuthology.orchestra.run.vm11.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T17:25:45.574 INFO:teuthology.orchestra.run.vm10.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T17:25:45.583 INFO:teuthology.orchestra.run.vm11.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T17:25:45.629 INFO:teuthology.orchestra.run.vm10.stderr:bash: line 1: ntpq: command not found 2026-03-09T17:25:45.630 INFO:teuthology.orchestra.run.vm11.stderr:bash: line 1: ntpq: command not found 2026-03-09T17:25:45.631 INFO:teuthology.orchestra.run.vm10.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T17:25:45.631 INFO:teuthology.orchestra.run.vm10.stdout:=============================================================================== 2026-03-09T17:25:45.633 INFO:teuthology.orchestra.run.vm11.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T17:25:45.633 INFO:teuthology.orchestra.run.vm11.stdout:=============================================================================== 2026-03-09T17:25:45.633 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-09T17:25:45.679 INFO:tasks.cephadm:Config: {'cephadm_branch': 'v17.2.0', 'cephadm_git_url': 'https://github.com/ceph/ceph', 'image': 'quay.io/ceph/ceph:v17.2.0', 'roleless': True, 'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_REFRESH_FAILED'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-09T17:25:45.679 INFO:tasks.cephadm:Cluster image is quay.io/ceph/ceph:v17.2.0 2026-03-09T17:25:45.679 INFO:tasks.cephadm:Cluster fsid is 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:25:45.679 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-09T17:25:45.679 INFO:tasks.cephadm:No mon roles; fabricating mons 2026-03-09T17:25:45.679 INFO:tasks.cephadm:Monitor IPs: {'mon.vm10': '192.168.123.110', 'mon.vm11': '192.168.123.111'} 2026-03-09T17:25:45.679 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-09T17:25:45.679 DEBUG:teuthology.orchestra.run.vm10:> sudo hostname $(hostname -s) 2026-03-09T17:25:45.714 DEBUG:teuthology.orchestra.run.vm11:> sudo hostname $(hostname -s) 2026-03-09T17:25:45.748 INFO:tasks.cephadm:Downloading cephadm (repo https://github.com/ceph/ceph ref v17.2.0)... 2026-03-09T17:25:45.748 DEBUG:teuthology.orchestra.run.vm10:> curl --silent https://raw.githubusercontent.com/ceph/ceph/v17.2.0/src/cephadm/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T17:25:46.019 INFO:teuthology.orchestra.run.vm10.stdout:-rw-r--r--. 1 ubuntu ubuntu 320521 Mar 9 17:25 /home/ubuntu/cephtest/cephadm 2026-03-09T17:25:46.020 DEBUG:teuthology.orchestra.run.vm11:> curl --silent https://raw.githubusercontent.com/ceph/ceph/v17.2.0/src/cephadm/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T17:25:46.106 INFO:teuthology.orchestra.run.vm11.stdout:-rw-r--r--. 1 ubuntu ubuntu 320521 Mar 9 17:25 /home/ubuntu/cephtest/cephadm 2026-03-09T17:25:46.107 DEBUG:teuthology.orchestra.run.vm10:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T17:25:46.134 DEBUG:teuthology.orchestra.run.vm11:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T17:25:46.159 INFO:tasks.cephadm:Pulling image quay.io/ceph/ceph:v17.2.0 on all hosts... 2026-03-09T17:25:46.159 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 pull 2026-03-09T17:25:46.176 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 pull 2026-03-09T17:25:46.394 INFO:teuthology.orchestra.run.vm11.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-09T17:25:46.396 INFO:teuthology.orchestra.run.vm10.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-09T17:26:08.424 INFO:teuthology.orchestra.run.vm11.stdout:{ 2026-03-09T17:26:08.424 INFO:teuthology.orchestra.run.vm11.stdout: "ceph_version": "ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable)", 2026-03-09T17:26:08.424 INFO:teuthology.orchestra.run.vm11.stdout: "image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", 2026-03-09T17:26:08.424 INFO:teuthology.orchestra.run.vm11.stdout: "repo_digests": [ 2026-03-09T17:26:08.424 INFO:teuthology.orchestra.run.vm11.stdout: "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", 2026-03-09T17:26:08.424 INFO:teuthology.orchestra.run.vm11.stdout: "quay.io/ceph/ceph@sha256:cb4d698cb769b6aba05bf6ef04f41a7fe694160140347576e13bd9348514b667" 2026-03-09T17:26:08.424 INFO:teuthology.orchestra.run.vm11.stdout: ] 2026-03-09T17:26:08.424 INFO:teuthology.orchestra.run.vm11.stdout:} 2026-03-09T17:26:10.438 INFO:teuthology.orchestra.run.vm10.stdout:{ 2026-03-09T17:26:10.438 INFO:teuthology.orchestra.run.vm10.stdout: "ceph_version": "ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable)", 2026-03-09T17:26:10.438 INFO:teuthology.orchestra.run.vm10.stdout: "image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", 2026-03-09T17:26:10.438 INFO:teuthology.orchestra.run.vm10.stdout: "repo_digests": [ 2026-03-09T17:26:10.438 INFO:teuthology.orchestra.run.vm10.stdout: "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", 2026-03-09T17:26:10.438 INFO:teuthology.orchestra.run.vm10.stdout: "quay.io/ceph/ceph@sha256:cb4d698cb769b6aba05bf6ef04f41a7fe694160140347576e13bd9348514b667" 2026-03-09T17:26:10.438 INFO:teuthology.orchestra.run.vm10.stdout: ] 2026-03-09T17:26:10.438 INFO:teuthology.orchestra.run.vm10.stdout:} 2026-03-09T17:26:10.476 DEBUG:teuthology.orchestra.run.vm10:> sudo mkdir -p /etc/ceph 2026-03-09T17:26:10.516 DEBUG:teuthology.orchestra.run.vm11:> sudo mkdir -p /etc/ceph 2026-03-09T17:26:10.555 DEBUG:teuthology.orchestra.run.vm10:> sudo chmod 777 /etc/ceph 2026-03-09T17:26:10.593 DEBUG:teuthology.orchestra.run.vm11:> sudo chmod 777 /etc/ceph 2026-03-09T17:26:10.633 INFO:tasks.cephadm:Writing seed config... 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-09T17:26:10.633 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-09T17:26:10.633 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:26:10.633 DEBUG:teuthology.orchestra.run.vm10:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-09T17:26:10.656 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 01c7db6a-1bdd-11f1-ac39-b134231210ea [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-09T17:26:10.657 DEBUG:teuthology.orchestra.run.vm10:mon.vm10> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@mon.vm10.service 2026-03-09T17:26:10.700 INFO:tasks.cephadm:Bootstrapping... 2026-03-09T17:26:10.700 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 -v bootstrap --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-ip 192.168.123.110 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:26:10.875 INFO:teuthology.orchestra.run.vm10.stderr:-------------------------------------------------------------------------------- 2026-03-09T17:26:10.875 INFO:teuthology.orchestra.run.vm10.stderr:cephadm ['--image', 'quay.io/ceph/ceph:v17.2.0', '-v', 'bootstrap', '--fsid', '01c7db6a-1bdd-11f1-ac39-b134231210ea', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-ip', '192.168.123.110', '--skip-admin-label'] 2026-03-09T17:26:10.904 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: 5.8.0 2026-03-09T17:26:10.907 INFO:teuthology.orchestra.run.vm10.stderr:Verifying podman|docker is present... 2026-03-09T17:26:10.927 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: 5.8.0 2026-03-09T17:26:10.931 INFO:teuthology.orchestra.run.vm10.stderr:Verifying lvm2 is present... 2026-03-09T17:26:10.931 INFO:teuthology.orchestra.run.vm10.stderr:Verifying time synchronization is in place... 2026-03-09T17:26:10.937 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T17:26:10.945 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: inactive 2026-03-09T17:26:10.953 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: enabled 2026-03-09T17:26:10.965 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: active 2026-03-09T17:26:10.966 INFO:teuthology.orchestra.run.vm10.stderr:Unit chronyd.service is enabled and running 2026-03-09T17:26:10.966 INFO:teuthology.orchestra.run.vm10.stderr:Repeating the final host check... 2026-03-09T17:26:10.985 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: 5.8.0 2026-03-09T17:26:10.987 INFO:teuthology.orchestra.run.vm10.stderr:podman (/bin/podman) version 5.8.0 is present 2026-03-09T17:26:10.987 INFO:teuthology.orchestra.run.vm10.stderr:systemctl is present 2026-03-09T17:26:10.987 INFO:teuthology.orchestra.run.vm10.stderr:lvcreate is present 2026-03-09T17:26:10.994 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T17:26:11.000 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: inactive 2026-03-09T17:26:11.012 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: enabled 2026-03-09T17:26:11.019 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: active 2026-03-09T17:26:11.020 INFO:teuthology.orchestra.run.vm10.stderr:Unit chronyd.service is enabled and running 2026-03-09T17:26:11.020 INFO:teuthology.orchestra.run.vm10.stderr:Host looks OK 2026-03-09T17:26:11.020 INFO:teuthology.orchestra.run.vm10.stderr:Cluster fsid: 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:11.020 INFO:teuthology.orchestra.run.vm10.stderr:Acquiring lock 140151913598832 on /run/cephadm/01c7db6a-1bdd-11f1-ac39-b134231210ea.lock 2026-03-09T17:26:11.020 INFO:teuthology.orchestra.run.vm10.stderr:Lock 140151913598832 acquired on /run/cephadm/01c7db6a-1bdd-11f1-ac39-b134231210ea.lock 2026-03-09T17:26:11.020 INFO:teuthology.orchestra.run.vm10.stderr:Verifying IP 192.168.123.110 port 3300 ... 2026-03-09T17:26:11.021 INFO:teuthology.orchestra.run.vm10.stderr:Verifying IP 192.168.123.110 port 6789 ... 2026-03-09T17:26:11.021 INFO:teuthology.orchestra.run.vm10.stderr:Base mon IP is 192.168.123.110, final addrv is [v2:192.168.123.110:3300,v1:192.168.123.110:6789] 2026-03-09T17:26:11.025 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.110 metric 100 2026-03-09T17:26:11.025 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.110 metric 100 2026-03-09T17:26:11.029 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: ::1 dev lo proto kernel metric 256 pref medium 2026-03-09T17:26:11.029 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-09T17:26:11.032 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-09T17:26:11.032 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: inet6 ::1/128 scope host 2026-03-09T17:26:11.032 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: valid_lft forever preferred_lft forever 2026-03-09T17:26:11.032 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: 2: eth0: mtu 1500 state UP qlen 1000 2026-03-09T17:26:11.032 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: inet6 fe80::5055:ff:fe00:a/64 scope link noprefixroute 2026-03-09T17:26:11.032 INFO:teuthology.orchestra.run.vm10.stderr:/sbin/ip: valid_lft forever preferred_lft forever 2026-03-09T17:26:11.033 INFO:teuthology.orchestra.run.vm10.stderr:Mon IP `192.168.123.110` is in CIDR network `192.168.123.0/24` 2026-03-09T17:26:11.033 INFO:teuthology.orchestra.run.vm10.stderr:- internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-09T17:26:11.033 INFO:teuthology.orchestra.run.vm10.stderr:Pulling container image quay.io/ceph/ceph:v17.2.0... 2026-03-09T17:26:11.055 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Trying to pull quay.io/ceph/ceph:v17.2.0... 2026-03-09T17:26:12.246 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Getting image source signatures 2026-03-09T17:26:12.246 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Copying blob sha256:33ca8fff7868c4dc0c11e09bca97c720eb9cfbab7221216754367dd8de70388a 2026-03-09T17:26:12.246 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Copying blob sha256:89b4a75bc2d8500f15463747507c9623df43886c134463e7f0527e70900e7a7b 2026-03-09T17:26:12.246 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Copying blob sha256:a70843738bb77e1ab9c1f85969ebdfa55f178e746be081d1cb4f94011f69eb7c 2026-03-09T17:26:12.246 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Copying blob sha256:c32ab78b488d0c72f64eded765c0cf6b5bf2c75dab66cb62a9d367fa6ec42513 2026-03-09T17:26:12.246 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Copying blob sha256:599d07cb321ff0a3c82224e1138fc685793fa69b93ed5780415751a5f7e4b8c2 2026-03-09T17:26:12.246 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Copying config sha256:e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9 2026-03-09T17:26:12.247 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: Writing manifest to image destination 2026-03-09T17:26:12.253 INFO:teuthology.orchestra.run.vm10.stderr:/bin/podman: e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9 2026-03-09T17:26:12.416 INFO:teuthology.orchestra.run.vm10.stderr:ceph: ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable) 2026-03-09T17:26:12.463 INFO:teuthology.orchestra.run.vm10.stderr:Ceph version: ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable) 2026-03-09T17:26:12.463 INFO:teuthology.orchestra.run.vm10.stderr:Extracting ceph user uid/gid from container image... 2026-03-09T17:26:12.542 INFO:teuthology.orchestra.run.vm10.stderr:stat: 167 167 2026-03-09T17:26:12.576 INFO:teuthology.orchestra.run.vm10.stderr:Creating initial keys... 2026-03-09T17:26:12.677 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph-authtool: AQC0Aq9pSpZQKBAA0oEXtJPh9FBaTOxCGFr1Lw== 2026-03-09T17:26:12.797 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph-authtool: AQC0Aq9p5zaCLxAAiXvvDgS1wYAdsh8vZ+ecaQ== 2026-03-09T17:26:12.916 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph-authtool: AQC0Aq9p7haVNhAAkPBAwccUdy6Yhn3yF3/4bg== 2026-03-09T17:26:12.938 INFO:teuthology.orchestra.run.vm10.stderr:Creating initial monmap... 2026-03-09T17:26:13.056 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T17:26:13.056 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/monmaptool: setting min_mon_release = octopus 2026-03-09T17:26:13.056 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: set fsid to 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:13.056 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/monmaptool: /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T17:26:13.073 INFO:teuthology.orchestra.run.vm10.stderr:monmaptool for vm10 [v2:192.168.123.110:3300,v1:192.168.123.110:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T17:26:13.073 INFO:teuthology.orchestra.run.vm10.stderr:setting min_mon_release = octopus 2026-03-09T17:26:13.073 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/monmaptool: set fsid to 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:13.073 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T17:26:13.073 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:13.073 INFO:teuthology.orchestra.run.vm10.stderr:Creating mon... 2026-03-09T17:26:13.250 INFO:teuthology.orchestra.run.vm10.stderr:create mon.vm10 on 2026-03-09T17:26:13.488 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-09T17:26:13.662 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea.target → /etc/systemd/system/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea.target. 2026-03-09T17:26:13.662 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Created symlink /etc/systemd/system/ceph.target.wants/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea.target → /etc/systemd/system/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea.target. 2026-03-09T17:26:14.037 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Failed to reset failed state of unit ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@mon.vm10.service: Unit ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@mon.vm10.service not loaded. 2026-03-09T17:26:14.052 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Created symlink /etc/systemd/system/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea.target.wants/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@mon.vm10.service → /etc/systemd/system/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@.service. 2026-03-09T17:26:15.171 INFO:teuthology.orchestra.run.vm10.stderr:firewalld does not appear to be present 2026-03-09T17:26:15.171 INFO:teuthology.orchestra.run.vm10.stderr:Not possible to enable service . firewalld.service is not available 2026-03-09T17:26:15.171 INFO:teuthology.orchestra.run.vm10.stderr:Waiting for mon to start... 2026-03-09T17:26:15.171 INFO:teuthology.orchestra.run.vm10.stderr:Waiting for mon... 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: cluster: 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: id: 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: health: HEALTH_OK 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: services: 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mon: 1 daemons, quorum vm10 (age 0.175175s) 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mgr: no daemons active 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd: 0 osds: 0 up, 0 in 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: data: 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: pools: 0 pools, 0 pgs 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: objects: 0 objects, 0 B 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: usage: 0 B used, 0 B / 0 B avail 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: pgs: 2026-03-09T17:26:15.403 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:15.443 INFO:teuthology.orchestra.run.vm10.stderr:mon is available 2026-03-09T17:26:15.443 INFO:teuthology.orchestra.run.vm10.stderr:Assimilating anything we can from ceph.conf... 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: [global] 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: fsid = 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mon_host = [v2:192.168.123.110:3300,v1:192.168.123.110:6789] 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mon_osd_allow_pg_remap = true 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mon_osd_allow_primary_affinity = true 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mon_warn_on_no_sortbitwise = false 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd_crush_chooseleaf_type = 0 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: [mgr] 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mgr/telemetry/nag = false 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: [osd] 2026-03-09T17:26:15.661 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd_map_max_advance = 10 2026-03-09T17:26:15.662 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd_mclock_iops_capacity_threshold_hdd = 49000 2026-03-09T17:26:15.662 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd_sloppy_crc = true 2026-03-09T17:26:15.728 INFO:teuthology.orchestra.run.vm10.stderr:Generating new minimal ceph.conf... 2026-03-09T17:26:15.985 INFO:teuthology.orchestra.run.vm10.stderr:Restarting the monitor... 2026-03-09T17:26:16.390 INFO:teuthology.orchestra.run.vm10.stderr:Setting mon public_network to 192.168.123.0/24 2026-03-09T17:26:16.685 INFO:teuthology.orchestra.run.vm10.stderr:Wrote config to /etc/ceph/ceph.conf 2026-03-09T17:26:16.685 INFO:teuthology.orchestra.run.vm10.stderr:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:26:16.685 INFO:teuthology.orchestra.run.vm10.stderr:Creating mgr... 2026-03-09T17:26:16.685 INFO:teuthology.orchestra.run.vm10.stderr:Verifying port 9283 ... 2026-03-09T17:26:16.890 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Failed to reset failed state of unit ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@mgr.vm10.qompzp.service: Unit ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@mgr.vm10.qompzp.service not loaded. 2026-03-09T17:26:16.899 INFO:teuthology.orchestra.run.vm10.stderr:systemctl: Created symlink /etc/systemd/system/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea.target.wants/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@mgr.vm10.qompzp.service → /etc/systemd/system/ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@.service. 2026-03-09T17:26:17.265 INFO:teuthology.orchestra.run.vm10.stderr:firewalld does not appear to be present 2026-03-09T17:26:17.265 INFO:teuthology.orchestra.run.vm10.stderr:Not possible to enable service . firewalld.service is not available 2026-03-09T17:26:17.265 INFO:teuthology.orchestra.run.vm10.stderr:firewalld does not appear to be present 2026-03-09T17:26:17.265 INFO:teuthology.orchestra.run.vm10.stderr:Not possible to open ports <[9283]>. firewalld.service is not available 2026-03-09T17:26:17.265 INFO:teuthology.orchestra.run.vm10.stderr:Waiting for mgr to start... 2026-03-09T17:26:17.265 INFO:teuthology.orchestra.run.vm10.stderr:Waiting for mgr... 2026-03-09T17:26:17.541 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:17.541 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: { 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "fsid": "01c7db6a-1bdd-11f1-ac39-b134231210ea", 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "health": { 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "checks": {}, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mutes": [] 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum": [ 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 0 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "vm10" 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum_age": 1, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "monmap": { 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osdmap": { 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "pgmap": { 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "fsmap": { 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "available": false, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-09T17:26:17.542 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "modules": [ 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "iostat", 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "nfs", 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "restful" 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "services": {} 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "servicemap": { 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "modified": "2026-03-09T17:26:15.226965+0000", 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "services": {} 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-09T17:26:17.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: } 2026-03-09T17:26:17.576 INFO:teuthology.orchestra.run.vm10.stderr:mgr not available, waiting (1/15)... 2026-03-09T17:26:19.840 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: { 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "fsid": "01c7db6a-1bdd-11f1-ac39-b134231210ea", 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "health": { 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "checks": {}, 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mutes": [] 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum": [ 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 0 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "vm10" 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum_age": 3, 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "monmap": { 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-09T17:26:19.841 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osdmap": { 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "pgmap": { 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "fsmap": { 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "available": false, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "modules": [ 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "iostat", 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "nfs", 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "restful" 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "services": {} 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "servicemap": { 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "modified": "2026-03-09T17:26:15.226965+0000", 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "services": {} 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:19.842 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-09T17:26:19.843 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: } 2026-03-09T17:26:19.882 INFO:teuthology.orchestra.run.vm10.stderr:mgr not available, waiting (2/15)... 2026-03-09T17:26:22.193 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: { 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "fsid": "01c7db6a-1bdd-11f1-ac39-b134231210ea", 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "health": { 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "checks": {}, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mutes": [] 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum": [ 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 0 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "vm10" 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum_age": 5, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "monmap": { 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osdmap": { 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-09T17:26:22.194 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "pgmap": { 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "fsmap": { 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "available": false, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "modules": [ 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "iostat", 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "nfs", 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "restful" 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "services": {} 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "servicemap": { 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "modified": "2026-03-09T17:26:15.226965+0000", 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "services": {} 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-09T17:26:22.195 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: } 2026-03-09T17:26:22.240 INFO:teuthology.orchestra.run.vm10.stderr:mgr not available, waiting (3/15)... 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: { 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "fsid": "01c7db6a-1bdd-11f1-ac39-b134231210ea", 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "health": { 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "status": "HEALTH_OK", 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "checks": {}, 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mutes": [] 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:24.559 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "election_epoch": 5, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum": [ 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 0 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum_names": [ 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "vm10" 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "quorum_age": 8, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "monmap": { 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "min_mon_release_name": "quincy", 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_mons": 1 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osdmap": { 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_osds": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_up_osds": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osd_up_since": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_in_osds": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "osd_in_since": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_remapped_pgs": 0 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "pgmap": { 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "pgs_by_state": [], 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_pgs": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_pools": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_objects": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "data_bytes": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_used": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_avail": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "bytes_total": 0 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "fsmap": { 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "by_rank": [], 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "up:standby": 0 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mgrmap": { 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "available": true, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_standbys": 0, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "modules": [ 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "iostat", 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "nfs", 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "restful" 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ], 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "services": {} 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "servicemap": { 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 1, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "modified": "2026-03-09T17:26:15.226965+0000", 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "services": {} 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: }, 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "progress_events": {} 2026-03-09T17:26:24.560 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: } 2026-03-09T17:26:24.745 INFO:teuthology.orchestra.run.vm10.stderr:mgr is available 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: [global] 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: fsid = 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mon_osd_allow_pg_remap = true 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mon_osd_allow_primary_affinity = true 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mon_warn_on_no_sortbitwise = false 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd_crush_chooseleaf_type = 0 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: [mgr] 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: mgr/telemetry/nag = false 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: [osd] 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd_map_max_advance = 10 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd_mclock_iops_capacity_threshold_hdd = 49000 2026-03-09T17:26:25.014 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: osd_sloppy_crc = true 2026-03-09T17:26:25.050 INFO:teuthology.orchestra.run.vm10.stderr:Enabling cephadm module... 2026-03-09T17:26:26.347 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: { 2026-03-09T17:26:26.347 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 5, 2026-03-09T17:26:26.347 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "available": true, 2026-03-09T17:26:26.347 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "active_name": "vm10.qompzp", 2026-03-09T17:26:26.347 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_standby": 0 2026-03-09T17:26:26.347 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: } 2026-03-09T17:26:26.391 INFO:teuthology.orchestra.run.vm10.stderr:Waiting for the mgr to restart... 2026-03-09T17:26:26.391 INFO:teuthology.orchestra.run.vm10.stderr:Waiting for mgr epoch 5... 2026-03-09T17:26:31.783 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: { 2026-03-09T17:26:31.783 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mgrmap_epoch": 7, 2026-03-09T17:26:31.783 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "initialized": true 2026-03-09T17:26:31.783 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: } 2026-03-09T17:26:31.855 INFO:teuthology.orchestra.run.vm10.stderr:mgr epoch 5 is available 2026-03-09T17:26:31.855 INFO:teuthology.orchestra.run.vm10.stderr:Setting orchestrator backend to cephadm... 2026-03-09T17:26:32.543 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: value unchanged 2026-03-09T17:26:32.575 INFO:teuthology.orchestra.run.vm10.stderr:Generating ssh key... 2026-03-09T17:26:33.203 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDR+YdA2S2rWA0v4Z5bsyV81f1n1HQt6jdxyuLdjtXG67MTg4s0i0/y6uVcfAOHcTx7XT02LNB6ldRHQQgvb99jHowomIKhB/My5LMi5Mzf1KXYOwRDeP9DQdIF/PGwTYQR4sJdud2jpWzOb2s4aCF5CW2okb3CFfWAdsD9kFcjMVKzi1BBEXhp/l1mN6K2FSQ/MDnu3dfe0STR8n1wFyZN0DU6eh9hFLGcz3btyYezbNXyFIUsY6Kj3RD+kGb09rfy3zX5UcH1ySjnPW6z5CE/Dmiwfd/sjzvBCMQnwDw2HParmK2nVRC2aDtoz+7+lzlOwptr70/eFdIvY1KLZlkoMJg2y9bOkG1ewoS/1y0GNYbWmplFq7+5DttxBFYuFH3mY53FhA7uvQIplAuVDMTTOyYNvi7yDp/+EHq2cwzB6cW27YBhvzxC6nVozm4okH7KA32aiDQqNU+xANs7mDq7mOhRIf5QCqoQvrFyqmFtv9s6pJSjemLHDzo1KrLOWUs= ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:33.238 INFO:teuthology.orchestra.run.vm10.stderr:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-09T17:26:33.238 INFO:teuthology.orchestra.run.vm10.stderr:Adding key to root@localhost authorized_keys... 2026-03-09T17:26:33.239 INFO:teuthology.orchestra.run.vm10.stderr:Adding host vm10... 2026-03-09T17:26:34.258 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Added host 'vm10' with addr '192.168.123.110' 2026-03-09T17:26:34.293 INFO:teuthology.orchestra.run.vm10.stderr:Deploying mon service with default placement... 2026-03-09T17:26:34.635 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Scheduled mon update... 2026-03-09T17:26:34.677 INFO:teuthology.orchestra.run.vm10.stderr:Deploying mgr service with default placement... 2026-03-09T17:26:34.952 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Scheduled mgr update... 2026-03-09T17:26:35.022 INFO:teuthology.orchestra.run.vm10.stderr:Deploying crash service with default placement... 2026-03-09T17:26:35.320 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Scheduled crash update... 2026-03-09T17:26:35.358 INFO:teuthology.orchestra.run.vm10.stderr:Deploying prometheus service with default placement... 2026-03-09T17:26:35.632 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Scheduled prometheus update... 2026-03-09T17:26:35.665 INFO:teuthology.orchestra.run.vm10.stderr:Deploying grafana service with default placement... 2026-03-09T17:26:36.152 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Scheduled grafana update... 2026-03-09T17:26:36.218 INFO:teuthology.orchestra.run.vm10.stderr:Deploying node-exporter service with default placement... 2026-03-09T17:26:36.718 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Scheduled node-exporter update... 2026-03-09T17:26:36.773 INFO:teuthology.orchestra.run.vm10.stderr:Deploying alertmanager service with default placement... 2026-03-09T17:26:37.157 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Scheduled alertmanager update... 2026-03-09T17:26:38.218 INFO:teuthology.orchestra.run.vm10.stderr:Enabling the dashboard module... 2026-03-09T17:26:39.551 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: { 2026-03-09T17:26:39.551 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "epoch": 9, 2026-03-09T17:26:39.551 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "available": true, 2026-03-09T17:26:39.551 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "active_name": "vm10.qompzp", 2026-03-09T17:26:39.551 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "num_standby": 0 2026-03-09T17:26:39.551 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: } 2026-03-09T17:26:39.602 INFO:teuthology.orchestra.run.vm10.stderr:Waiting for the mgr to restart... 2026-03-09T17:26:39.602 INFO:teuthology.orchestra.run.vm10.stderr:Waiting for mgr epoch 9... 2026-03-09T17:26:45.065 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: { 2026-03-09T17:26:45.065 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "mgrmap_epoch": 11, 2026-03-09T17:26:45.065 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: "initialized": true 2026-03-09T17:26:45.065 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: } 2026-03-09T17:26:45.144 INFO:teuthology.orchestra.run.vm10.stderr:mgr epoch 9 is available 2026-03-09T17:26:45.144 INFO:teuthology.orchestra.run.vm10.stderr:Generating a dashboard self-signed certificate... 2026-03-09T17:26:45.569 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: Self-signed certificate created 2026-03-09T17:26:45.640 INFO:teuthology.orchestra.run.vm10.stderr:Creating initial admin user... 2026-03-09T17:26:46.213 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: {"username": "admin", "password": "$2b$12$zkVKih.OK7RQlaEHZX.pPe1XcAKPSKGFQrSu6KwEyNr465MWZssuy", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773077206, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-09T17:26:46.240 INFO:teuthology.orchestra.run.vm10.stderr:Fetching dashboard port number... 2026-03-09T17:26:46.551 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: 8443 2026-03-09T17:26:46.585 INFO:teuthology.orchestra.run.vm10.stderr:firewalld does not appear to be present 2026-03-09T17:26:46.586 INFO:teuthology.orchestra.run.vm10.stderr:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-09T17:26:46.588 INFO:teuthology.orchestra.run.vm10.stderr:Ceph Dashboard is now available at: 2026-03-09T17:26:46.588 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:46.588 INFO:teuthology.orchestra.run.vm10.stderr: URL: https://vm10.local:8443/ 2026-03-09T17:26:46.588 INFO:teuthology.orchestra.run.vm10.stderr: User: admin 2026-03-09T17:26:46.588 INFO:teuthology.orchestra.run.vm10.stderr: Password: wpga2r2j2w 2026-03-09T17:26:46.588 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:46.588 INFO:teuthology.orchestra.run.vm10.stderr:Enabling autotune for osd_memory_target 2026-03-09T17:26:47.245 INFO:teuthology.orchestra.run.vm10.stderr:/usr/bin/ceph: set mgr/dashboard/cluster/status 2026-03-09T17:26:47.274 INFO:teuthology.orchestra.run.vm10.stderr:You can access the Ceph CLI with: 2026-03-09T17:26:47.274 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:47.274 INFO:teuthology.orchestra.run.vm10.stderr: sudo /home/ubuntu/cephtest/cephadm shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-09T17:26:47.274 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr:Please consider enabling telemetry to help improve Ceph: 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr: ceph telemetry on 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr:For more information see: 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr: https://docs.ceph.com/docs/master/mgr/telemetry/ 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T17:26:47.275 INFO:teuthology.orchestra.run.vm10.stderr:Bootstrap complete. 2026-03-09T17:26:47.336 INFO:tasks.cephadm:Fetching config... 2026-03-09T17:26:47.336 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:26:47.336 DEBUG:teuthology.orchestra.run.vm10:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-09T17:26:47.360 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-09T17:26:47.360 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:26:47.360 DEBUG:teuthology.orchestra.run.vm10:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-09T17:26:47.439 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-09T17:26:47.439 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:26:47.439 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/keyring of=/dev/stdout 2026-03-09T17:26:47.529 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-09T17:26:47.530 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:26:47.530 DEBUG:teuthology.orchestra.run.vm10:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-09T17:26:47.602 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-09T17:26:47.602 DEBUG:teuthology.orchestra.run.vm10:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDR+YdA2S2rWA0v4Z5bsyV81f1n1HQt6jdxyuLdjtXG67MTg4s0i0/y6uVcfAOHcTx7XT02LNB6ldRHQQgvb99jHowomIKhB/My5LMi5Mzf1KXYOwRDeP9DQdIF/PGwTYQR4sJdud2jpWzOb2s4aCF5CW2okb3CFfWAdsD9kFcjMVKzi1BBEXhp/l1mN6K2FSQ/MDnu3dfe0STR8n1wFyZN0DU6eh9hFLGcz3btyYezbNXyFIUsY6Kj3RD+kGb09rfy3zX5UcH1ySjnPW6z5CE/Dmiwfd/sjzvBCMQnwDw2HParmK2nVRC2aDtoz+7+lzlOwptr70/eFdIvY1KLZlkoMJg2y9bOkG1ewoS/1y0GNYbWmplFq7+5DttxBFYuFH3mY53FhA7uvQIplAuVDMTTOyYNvi7yDp/+EHq2cwzB6cW27YBhvzxC6nVozm4okH7KA32aiDQqNU+xANs7mDq7mOhRIf5QCqoQvrFyqmFtv9s6pJSjemLHDzo1KrLOWUs= ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T17:26:47.702 INFO:teuthology.orchestra.run.vm10.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDR+YdA2S2rWA0v4Z5bsyV81f1n1HQt6jdxyuLdjtXG67MTg4s0i0/y6uVcfAOHcTx7XT02LNB6ldRHQQgvb99jHowomIKhB/My5LMi5Mzf1KXYOwRDeP9DQdIF/PGwTYQR4sJdud2jpWzOb2s4aCF5CW2okb3CFfWAdsD9kFcjMVKzi1BBEXhp/l1mN6K2FSQ/MDnu3dfe0STR8n1wFyZN0DU6eh9hFLGcz3btyYezbNXyFIUsY6Kj3RD+kGb09rfy3zX5UcH1ySjnPW6z5CE/Dmiwfd/sjzvBCMQnwDw2HParmK2nVRC2aDtoz+7+lzlOwptr70/eFdIvY1KLZlkoMJg2y9bOkG1ewoS/1y0GNYbWmplFq7+5DttxBFYuFH3mY53FhA7uvQIplAuVDMTTOyYNvi7yDp/+EHq2cwzB6cW27YBhvzxC6nVozm4okH7KA32aiDQqNU+xANs7mDq7mOhRIf5QCqoQvrFyqmFtv9s6pJSjemLHDzo1KrLOWUs= ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:47.723 DEBUG:teuthology.orchestra.run.vm11:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDR+YdA2S2rWA0v4Z5bsyV81f1n1HQt6jdxyuLdjtXG67MTg4s0i0/y6uVcfAOHcTx7XT02LNB6ldRHQQgvb99jHowomIKhB/My5LMi5Mzf1KXYOwRDeP9DQdIF/PGwTYQR4sJdud2jpWzOb2s4aCF5CW2okb3CFfWAdsD9kFcjMVKzi1BBEXhp/l1mN6K2FSQ/MDnu3dfe0STR8n1wFyZN0DU6eh9hFLGcz3btyYezbNXyFIUsY6Kj3RD+kGb09rfy3zX5UcH1ySjnPW6z5CE/Dmiwfd/sjzvBCMQnwDw2HParmK2nVRC2aDtoz+7+lzlOwptr70/eFdIvY1KLZlkoMJg2y9bOkG1ewoS/1y0GNYbWmplFq7+5DttxBFYuFH3mY53FhA7uvQIplAuVDMTTOyYNvi7yDp/+EHq2cwzB6cW27YBhvzxC6nVozm4okH7KA32aiDQqNU+xANs7mDq7mOhRIf5QCqoQvrFyqmFtv9s6pJSjemLHDzo1KrLOWUs= ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T17:26:47.765 INFO:teuthology.orchestra.run.vm11.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDR+YdA2S2rWA0v4Z5bsyV81f1n1HQt6jdxyuLdjtXG67MTg4s0i0/y6uVcfAOHcTx7XT02LNB6ldRHQQgvb99jHowomIKhB/My5LMi5Mzf1KXYOwRDeP9DQdIF/PGwTYQR4sJdud2jpWzOb2s4aCF5CW2okb3CFfWAdsD9kFcjMVKzi1BBEXhp/l1mN6K2FSQ/MDnu3dfe0STR8n1wFyZN0DU6eh9hFLGcz3btyYezbNXyFIUsY6Kj3RD+kGb09rfy3zX5UcH1ySjnPW6z5CE/Dmiwfd/sjzvBCMQnwDw2HParmK2nVRC2aDtoz+7+lzlOwptr70/eFdIvY1KLZlkoMJg2y9bOkG1ewoS/1y0GNYbWmplFq7+5DttxBFYuFH3mY53FhA7uvQIplAuVDMTTOyYNvi7yDp/+EHq2cwzB6cW27YBhvzxC6nVozm4okH7KA32aiDQqNU+xANs7mDq7mOhRIf5QCqoQvrFyqmFtv9s6pJSjemLHDzo1KrLOWUs= ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:26:47.777 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-09T17:26:48.951 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-09T17:26:48.951 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-09T17:26:49.908 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm11 2026-03-09T17:26:49.908 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T17:26:49.908 DEBUG:teuthology.orchestra.run.vm11:> dd of=/etc/ceph/ceph.conf 2026-03-09T17:26:49.925 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T17:26:49.925 DEBUG:teuthology.orchestra.run.vm11:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:26:49.983 INFO:tasks.cephadm:Adding host vm11 to orchestrator... 2026-03-09T17:26:49.983 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch host add vm11 2026-03-09T17:26:51.508 INFO:teuthology.orchestra.run.vm10.stdout:Added host 'vm11' with addr '192.168.123.111' 2026-03-09T17:26:51.593 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch host ls --format=json 2026-03-09T17:26:52.767 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:26:52.767 INFO:teuthology.orchestra.run.vm10.stdout:[{"addr": "192.168.123.110", "hostname": "vm10", "labels": [], "status": ""}, {"addr": "192.168.123.111", "hostname": "vm11", "labels": [], "status": ""}] 2026-03-09T17:26:52.842 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-09T17:26:52.842 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd crush tunables default 2026-03-09T17:26:53.808 INFO:teuthology.orchestra.run.vm10.stderr:adjusted tunables profile to default 2026-03-09T17:26:54.060 INFO:tasks.cephadm:Adding mon.vm10 on vm10 2026-03-09T17:26:54.060 INFO:tasks.cephadm:Adding mon.vm11 on vm11 2026-03-09T17:26:54.060 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch apply mon '2;vm10:192.168.123.110=vm10;vm11:192.168.123.111=vm11' 2026-03-09T17:26:54.600 INFO:teuthology.orchestra.run.vm11.stdout:Scheduled mon update... 2026-03-09T17:26:54.667 DEBUG:teuthology.orchestra.run.vm11:mon.vm11> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@mon.vm11.service 2026-03-09T17:26:54.668 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:26:54.668 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:26:55.255 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:26:55.256 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:26:55.258 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:26:56.324 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:26:56.325 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:26:56.932 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:26:56.933 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:26:56.935 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:26:58.005 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:26:58.005 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:26:58.556 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:26:58.556 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:26:58.558 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:26:59.605 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:26:59.606 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:00.126 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:00.126 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:00.129 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:01.182 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:01.182 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:01.718 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:01.718 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:01.720 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:02.775 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:02.775 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:03.301 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:03.301 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:03.303 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:04.372 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:04.372 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:04.900 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:04.900 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:04.902 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:05.970 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:05.970 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:06.480 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:06.480 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:06.482 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:07.543 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:07.543 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:08.088 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:08.089 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:08.093 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:08 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/2656941959' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:09.162 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:09.162 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:09.206 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:09 vm10 ceph-mon[53712]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:09.685 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:09.685 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:09.688 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:10.210 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:10 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/4176796618' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:10.737 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:10.737 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:11.218 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:11 vm10 ceph-mon[53712]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:11.273 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:11.274 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:11.276 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:12.295 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:12 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/1927791854' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:12.328 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:12.329 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:12.977 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:12.977 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:12.978 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:13.299 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:13 vm10 ceph-mon[53712]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:13.299 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:13 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/2626721709' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:14.021 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:14.022 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:14.614 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:14.614 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:14.617 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:15.135 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:15 vm10 ceph-mon[53712]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:15.135 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:15 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/817636894' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:15.663 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:15.663 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:16.189 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:16.189 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:16.192 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:17.271 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:17.272 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:17.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:17 vm10 ceph-mon[53712]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:17.446 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:17 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/823232351' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:18.088 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:18.089 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:18.091 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:18.461 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:18 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/3705650737' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:19.163 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:19.163 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:19.463 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:19 vm10 ceph-mon[53712]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:19.463 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:19 vm10 ceph-mon[53712]: from='mgr.14162 192.168.123.110:0/3117200532' entity='mgr.vm10.qompzp' 2026-03-09T17:27:19.463 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:19 vm10 ceph-mon[53712]: from='mgr.14162 192.168.123.110:0/3117200532' entity='mgr.vm10.qompzp' 2026-03-09T17:27:19.665 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:19.665 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:19.668 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:20.473 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:20 vm10 ceph-mon[53712]: Deploying daemon node-exporter.vm10 on vm10 2026-03-09T17:27:20.473 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:20 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/2305568967' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:20.715 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:20.715 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:21.206 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:21.206 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:21.208 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:21.482 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:21 vm10 ceph-mon[53712]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:21.482 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:21 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/73633211' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:22.272 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:22.272 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:23.259 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:23.259 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:23.262 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:23.497 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:23 vm10 ceph-mon[53712]: from='mgr.14162 192.168.123.110:0/3117200532' entity='mgr.vm10.qompzp' 2026-03-09T17:27:23.497 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:23 vm10 ceph-mon[53712]: from='mgr.14162 192.168.123.110:0/3117200532' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-09T17:27:23.497 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:23 vm10 ceph-mon[53712]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:24.313 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:24.313 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:24.503 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:24 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/377565235' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:24.503 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:24 vm10 ceph-mon[53712]: from='mgr.14162 192.168.123.110:0/3117200532' entity='mgr.vm10.qompzp' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-09T17:27:24.503 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:24 vm10 ceph-mon[53712]: mgrmap e14: vm10.qompzp(active, since 39s) 2026-03-09T17:27:24.872 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:24.872 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:24.873 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:25.510 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:25 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/2476099908' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:25.938 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:25.939 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:26.459 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:26.459 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:26.461 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:26.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:26 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/1071209209' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:27.515 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:27.515 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:28.038 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:28.038 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:28.040 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:28.075 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:28 vm10 ceph-mon[53712]: Active manager daemon vm10.qompzp restarted 2026-03-09T17:27:28.075 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:28 vm10 ceph-mon[53712]: Activating manager daemon vm10.qompzp 2026-03-09T17:27:28.075 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:28 vm10 ceph-mon[53712]: osdmap e5: 0 total, 0 up, 0 in 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: mgrmap e15: vm10.qompzp(active, starting, since 0.203717s) 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr metadata", "who": "vm10.qompzp", "id": "vm10.qompzp"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: Manager daemon vm10.qompzp is now available 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/1472527998' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:29.091 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:29.103 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:29.103 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:29.657 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:29.657 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:29.668 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:30.123 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:30 vm10 ceph-mon[53712]: mgrmap e16: vm10.qompzp(active, since 1.21182s) 2026-03-09T17:27:30.123 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:30 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/2841226290' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:30.123 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:30 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:30.123 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:30 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:30.123 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:30 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:30.737 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:30.737 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:31.131 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:31 vm10 ceph-mon[53712]: [09/Mar/2026:17:27:29] ENGINE Bus STARTING 2026-03-09T17:27:31.131 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:31 vm10 ceph-mon[53712]: [09/Mar/2026:17:27:29] ENGINE Serving on https://192.168.123.110:7150 2026-03-09T17:27:31.131 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:31 vm10 ceph-mon[53712]: [09/Mar/2026:17:27:29] ENGINE Bus STARTED 2026-03-09T17:27:31.131 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:31 vm10 ceph-mon[53712]: mgrmap e17: vm10.qompzp(active, since 2s) 2026-03-09T17:27:31.131 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:31 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:31.131 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:31 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:27:31.131 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:31 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:31.351 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:31.351 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:31.358 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: Updating vm10:/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm11", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm11", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: Deploying daemon crash.vm11 on vm11 2026-03-09T17:27:32.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:32 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/4222547725' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:32.427 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:32.427 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:33.149 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:33.149 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:26:13.054583Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T17:27:33.152 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T17:27:33.412 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:33.412 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm11.ohhvcs", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:27:33.412 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm11.ohhvcs", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: Deploying daemon mgr.vm11.ohhvcs on vm11 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/1841562050' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:33.413 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:33 vm10 ceph-mon[53712]: Deploying daemon mon.vm11 on vm11 2026-03-09T17:27:34.144 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 systemd[1]: Starting Ceph mon.vm11 for 01c7db6a-1bdd-11f1-ac39-b134231210ea... 2026-03-09T17:27:34.204 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-09T17:27:34.204 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mon dump -f json 2026-03-09T17:27:34.440 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 podman[53959]: 2026-03-09 17:27:34.144055852 +0000 UTC m=+0.023970941 container create 5d27ff0af313a59db77f889420bb370731351a3ac2f63eb2a767c144d9eef7e2 (image=quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a, name=ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-mon-vm11, GIT_COMMIT=b613db0f44179c0940781c1c7fe04e1acb7093ac, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image, com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, io.k8s.display-name=CentOS Stream 8, maintainer=Guillaume Abrioux , build-date=2022-05-03T08:36:31.336870, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-754, vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., ceph=True, com.redhat.component=centos-stream-container, io.openshift.expose-services=, name=centos-stream, release=754, GIT_BRANCH=HEAD, GIT_CLEAN=True, RELEASE=HEAD, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, vcs-type=git, CEPH_POINT_RELEASE=-17.2.0, GIT_REPO=https://github.com/ceph/ceph-container.git, com.redhat.license_terms=https://centos.org/legal/licensing-policy/, distribution-scope=public, vendor=Red Hat, Inc., version=8, io.buildah.version=1.19.8, io.openshift.tags=base centos centos-stream, architecture=x86_64) 2026-03-09T17:27:34.440 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 podman[53959]: 2026-03-09 17:27:34.188654393 +0000 UTC m=+0.068569492 container init 5d27ff0af313a59db77f889420bb370731351a3ac2f63eb2a767c144d9eef7e2 (image=quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a, name=ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-mon-vm11, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, distribution-scope=public, architecture=x86_64, GIT_COMMIT=b613db0f44179c0940781c1c7fe04e1acb7093ac, GIT_REPO=https://github.com/ceph/ceph-container.git, build-date=2022-05-03T08:36:31.336870, com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, ceph=True, RELEASE=HEAD, maintainer=Guillaume Abrioux , release=754, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-754, io.openshift.tags=base centos centos-stream, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., io.buildah.version=1.19.8, io.k8s.display-name=CentOS Stream 8, name=centos-stream, GIT_CLEAN=True, vendor=Red Hat, Inc., version=8, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image, vcs-type=git, com.redhat.component=centos-stream-container, CEPH_POINT_RELEASE=-17.2.0, io.openshift.expose-services=, GIT_BRANCH=HEAD, com.redhat.license_terms=https://centos.org/legal/licensing-policy/, vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb) 2026-03-09T17:27:34.440 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 podman[53959]: 2026-03-09 17:27:34.19575131 +0000 UTC m=+0.075666389 container start 5d27ff0af313a59db77f889420bb370731351a3ac2f63eb2a767c144d9eef7e2 (image=quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a, name=ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-mon-vm11, GIT_BRANCH=HEAD, com.redhat.license_terms=https://centos.org/legal/licensing-policy/, description=CentOS Stream is a continuously delivered distro that tracks just ahead of Red Hat Enterprise Linux development. This image takes the Red Hat UBI and layers on content from CentOS Stream, GIT_COMMIT=b613db0f44179c0940781c1c7fe04e1acb7093ac, vcs-type=git, io.k8s.display-name=CentOS Stream 8, CEPH_POINT_RELEASE=-17.2.0, GIT_REPO=https://github.com/ceph/ceph-container.git, io.openshift.expose-services=, build-date=2022-05-03T08:36:31.336870, name=centos-stream, url=https://access.redhat.com/containers/#/registry.access.redhat.com/ubi8/images/8.6-754, GIT_CLEAN=True, io.openshift.tags=base centos centos-stream, io.k8s.description=The Universal Base Image is designed and engineered to be the base layer for all of your containerized applications, middleware and utilities. This base image is freely redistributable, but Red Hat only supports Red Hat technologies through subscriptions for Red Hat products. This image is maintained by Red Hat and updated regularly., architecture=x86_64, release=754, vcs-ref=f1ee6e37554363ec55e0035aba1a693d3627fdeb, RELEASE=HEAD, vendor=Red Hat, Inc., io.buildah.version=1.19.8, ceph=True, com.redhat.build-host=cpt-1002.osbs.prod.upshift.rdu2.redhat.com, distribution-scope=public, version=8, summary=Provides a CentOS Stream container based on the Red Hat Universal Base Image, com.redhat.component=centos-stream-container, maintainer=Guillaume Abrioux ) 2026-03-09T17:27:34.440 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 bash[53959]: 5d27ff0af313a59db77f889420bb370731351a3ac2f63eb2a767c144d9eef7e2 2026-03-09T17:27:34.440 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 podman[53959]: 2026-03-09 17:27:34.130420292 +0000 UTC m=+0.010335392 image pull e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9 quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 systemd[1]: Started Ceph mon.vm11 for 01c7db6a-1bdd-11f1-ac39-b134231210ea. 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: set uid:gid to 167:167 (ceph:ceph) 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: ceph version 17.2.0 (43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable), process ceph-mon, pid 2 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: pidfile_write: ignore empty --pid-file 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: load: jerasure load: lrc 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: RocksDB version: 6.15.5 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Git sha rocksdb_build_git_sha:@0@ 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Compile date Apr 18 2022 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: DB SUMMARY 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: DB Session ID: U5WL5CUESWMKI7Q62EU0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: CURRENT file: CURRENT 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: IDENTITY file: IDENTITY 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: MANIFEST file: MANIFEST-000003 size: 57 Bytes 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: SST files in /var/lib/ceph/mon/ceph-vm11/store.db dir, Total Num: 0, files: 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm11/store.db: 000004.log size: 511 ; 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.error_if_exists: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.create_if_missing: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.paranoid_checks: 1 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.env: 0x5620f56c6860 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.fs: Posix File System 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.info_log: 0x5620f75dbee0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_file_opening_threads: 16 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.statistics: (nil) 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.use_fsync: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_log_file_size: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.keep_log_file_num: 1000 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.recycle_log_file_num: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.allow_fallocate: 1 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.allow_mmap_reads: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.allow_mmap_writes: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.use_direct_reads: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.create_missing_column_families: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.db_log_dir: 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.wal_dir: /var/lib/ceph/mon/ceph-vm11/store.db 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-09T17:27:34.441 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.advise_random_on_open: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.db_write_buffer_size: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.write_buffer_manager: 0x5620f76cc2a0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.new_table_reader_for_compaction_inputs: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.rate_limiter: (nil) 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.wal_recovery_mode: 2 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.enable_thread_tracking: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.enable_pipelined_write: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.unordered_write: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.row_cache: None 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.wal_filter: None 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.allow_ingest_behind: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.preserve_deletes: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.two_write_queues: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.manual_wal_flush: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.atomic_flush: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.log_readahead_size: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.best_efforts_recovery: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.allow_data_in_errors: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.db_host_id: __hostname__ 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_background_jobs: 2 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_background_compactions: -1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_subcompactions: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_total_wal_size: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_open_files: -1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bytes_per_sync: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_readahead_size: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_background_flushes: -1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Compression algorithms supported: 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: kZSTD supported: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: kXpressCompression supported: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: kLZ4HCCompression supported: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: kLZ4Compression supported: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: kBZip2Compression supported: 0 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: kZlibCompression supported: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: kSnappyCompression supported: 1 2026-03-09T17:27:34.442 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/version_set.cc:4725] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm11/store.db/MANIFEST-000003 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/column_family.cc:597] --------------- Options for column family [default]: 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.merge_operator: 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_filter: None 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_filter_factory: None 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.sst_partitioner_factory: None 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5620f75a7d00) 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: cache_index_and_filter_blocks: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: pin_top_level_index_and_filter: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: index_type: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: data_block_index_type: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: index_shortening: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: hash_index_allow_collision: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: checksum: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: no_block_cache: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: block_cache: 0x5620f7612170 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: block_cache_name: BinnedLRUCache 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: block_cache_options: 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: capacity : 536870912 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: num_shard_bits : 4 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: strict_capacity_limit : 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: high_pri_pool_ratio: 0.000 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: block_cache_compressed: (nil) 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: persistent_cache: (nil) 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: block_size: 4096 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: block_size_deviation: 10 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: block_restart_interval: 16 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: index_block_restart_interval: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: metadata_block_size: 4096 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: partition_filters: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: use_delta_encoding: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: filter_policy: rocksdb.BuiltinBloomFilter 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: whole_key_filtering: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: verify_compression: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: read_amp_bytes_per_bit: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: format_version: 4 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: enable_index_compression: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout: block_align: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.write_buffer_size: 33554432 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_write_buffer_number: 2 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compression: NoCompression 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bottommost_compression: Disabled 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.prefix_extractor: nullptr 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.num_levels: 7 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-09T17:27:34.443 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compression_opts.level: 32767 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compression_opts.strategy: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compression_opts.enabled: false 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.target_file_size_base: 67108864 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.arena_block_size: 4194304 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.rate_limit_delay_max_milliseconds: 100 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.disable_auto_compactions: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.table_properties_collectors: 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.inplace_update_support: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.bloom_locality: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.max_successive_merges: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.paranoid_file_checks: 0 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.force_consistency_checks: 1 2026-03-09T17:27:34.444 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.report_bg_io_stats: 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.ttl: 2592000 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.enable_blob_files: false 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.min_blob_size: 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.blob_file_size: 268435456 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/version_set.cc:4773] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm11/store.db/MANIFEST-000003 succeeded,manifest_file_number is 3, next_file_number is 5, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/version_set.cc:4782] Column family [default] (ID 0), log number is 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/version_set.cc:4083] Creating manifest 7 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773077254231941, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/db_impl/db_impl_open.cc:847] Recovering log #4 mode 2 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [table/block_based/filter_policy.cc:996] Using legacy Bloom filter with high (20) bits/key. Dramatic filter space and/or accuracy improvement is available with format_version>=5. 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773077254233584, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1540, "file_checksum": "", "file_checksum_func_name": "Unknown", "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "rocksdb.BuiltinBloomFilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; ", "creation_time": 1773077254, "oldest_key_time": 0, "file_creation_time": 0, "db_id": "ff1c0e47-3fb1-40d2-8dbe-13e0b1defb75", "db_session_id": "U5WL5CUESWMKI7Q62EU0"}} 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/version_set.cc:4083] Creating manifest 9 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773077254236087, "job": 1, "event": "recovery_finished"} 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [file/delete_scheduler.cc:73] Deleted file /var/lib/ceph/mon/ceph-vm11/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/db_impl/db_impl_open.cc:1701] SstFileManager instance 0x5620f75f8700 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: DB pointer 0x5620f766c000 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11 does not exist in monmap, will attempt to join an existing cluster 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: using public_addr v2:192.168.123.111:0/0 -> [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/db_impl/db_impl.cc:902] ------- DUMPING STATS ------- 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: rocksdb: [db/db_impl/db_impl.cc:903] 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ** DB Stats ** 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 MB, 0.00 MB/s 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ** Compaction Stats [default] ** 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: L0 1/0 1.50 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Sum 1/0 1.50 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ** Compaction Stats [default] ** 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.9 0.00 0.00 1 0.002 0 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Cumulative compaction: 0.00 GB write, 0.07 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Interval compaction: 0.00 GB write, 0.07 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ** Compaction Stats [default] ** 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: L0 1/0 1.50 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 2026-03-09T17:27:34.445 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Sum 1/0 1.50 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.9 0.00 0.00 1 0.002 0 0 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ** Compaction Stats [default] ** 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.9 0.00 0.00 1 0.002 0 0 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Cumulative compaction: 0.00 GB write, 0.07 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: starting mon.vm11 rank -1 at public addrs [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] at bind addrs [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon_data /var/lib/ceph/mon/ceph-vm11 fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(???) e0 preinit fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/73633211' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14162 192.168.123.110:0/3117200532' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14162 192.168.123.110:0/3117200532' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/377565235' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14162 192.168.123.110:0/3117200532' entity='mgr.vm10.qompzp' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mgrmap e14: vm10.qompzp(active, since 39s) 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/2476099908' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/1071209209' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: Active manager daemon vm10.qompzp restarted 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: Activating manager daemon vm10.qompzp 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: osdmap e5: 0 total, 0 up, 0 in 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mgrmap e15: vm10.qompzp(active, starting, since 0.203717s) 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr metadata", "who": "vm10.qompzp", "id": "vm10.qompzp"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: Manager daemon vm10.qompzp is now available 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/1472527998' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mgrmap e16: vm10.qompzp(active, since 1.21182s) 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/2841226290' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: [09/Mar/2026:17:27:29] ENGINE Bus STARTING 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: [09/Mar/2026:17:27:29] ENGINE Serving on https://192.168.123.110:7150 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: [09/Mar/2026:17:27:29] ENGINE Bus STARTED 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mgrmap e17: vm10.qompzp(active, since 2s) 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.446 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: Updating vm10:/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm11", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm11", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: Deploying daemon crash.vm11 on vm11 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/4222547725' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm11.ohhvcs", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm11.ohhvcs", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: Deploying daemon mgr.vm11.ohhvcs on vm11 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/1841562050' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: Deploying daemon mon.vm11 on vm11 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).mds e1 new map 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).mds e1 print_map 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout: e1 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout: legacy client fscid: -1 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout: 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout: No filesystems configured 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e5 e5: 0 total, 0 up, 0 in 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e5 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing).paxosservice(auth 1..6) refresh upgraded, format 0 -> 3 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta expand map: {default=false} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta from 'false' to 'false' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta expanded map: {default=false} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta expand map: {default=info} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta from 'info' to 'info' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta expanded map: {default=info} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta expand map: {default=daemon} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta from 'daemon' to 'daemon' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta expanded map: {default=daemon} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta expand map: {default=debug} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta from 'debug' to 'debug' 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: expand_channel_meta expanded map: {default=debug} 2026-03-09T17:27:34.447 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:34 vm11 ceph-mon[53973]: mon.vm11@-1(synchronizing) e1 handle_conf_change mon_allow_pool_delete,mon_cluster_log_to_file 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: Deploying daemon node-exporter.vm11 on vm11 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: mon.vm10 calling monitor election 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: mon.vm11 calling monitor election 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.? 192.168.123.111:0/360025992' entity='mgr.vm11.ohhvcs' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm11.ohhvcs/crt"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: mon.vm10 is new leader, mons vm10,vm11 in quorum (ranks 0,1) 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: monmap e2: 2 mons at {vm10=[v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0],vm11=[v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0]} 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: fsmap 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: osdmap e5: 0 total, 0 up, 0 in 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: mgrmap e17: vm10.qompzp(active, since 11s) 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: Standby manager daemon vm11.ohhvcs started 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.? 192.168.123.111:0/360025992' entity='mgr.vm11.ohhvcs' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: overall HEALTH_OK 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.? 192.168.123.111:0/360025992' entity='mgr.vm11.ohhvcs' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm11.ohhvcs/key"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.? 192.168.123.111:0/360025992' entity='mgr.vm11.ohhvcs' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: Deploying daemon node-exporter.vm11 on vm11 2026-03-09T17:27:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T17:27:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: mon.vm10 calling monitor election 2026-03-09T17:27:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: mon.vm11 calling monitor election 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.? 192.168.123.111:0/360025992' entity='mgr.vm11.ohhvcs' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm11.ohhvcs/crt"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: mon.vm10 is new leader, mons vm10,vm11 in quorum (ranks 0,1) 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: monmap e2: 2 mons at {vm10=[v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0],vm11=[v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0]} 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: fsmap 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: osdmap e5: 0 total, 0 up, 0 in 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: mgrmap e17: vm10.qompzp(active, since 11s) 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: Standby manager daemon vm11.ohhvcs started 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.? 192.168.123.111:0/360025992' entity='mgr.vm11.ohhvcs' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: overall HEALTH_OK 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.? 192.168.123.111:0/360025992' entity='mgr.vm11.ohhvcs' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm11.ohhvcs/key"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.? 192.168.123.111:0/360025992' entity='mgr.vm11.ohhvcs' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:39.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:40.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:40 vm10 ceph-mon[53712]: Deploying daemon prometheus.vm10 on vm10 2026-03-09T17:27:40.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:40 vm10 ceph-mon[53712]: mgrmap e18: vm10.qompzp(active, since 11s), standbys: vm11.ohhvcs 2026-03-09T17:27:40.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr metadata", "who": "vm11.ohhvcs", "id": "vm11.ohhvcs"}]: dispatch 2026-03-09T17:27:40.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:40 vm11 ceph-mon[53973]: Deploying daemon prometheus.vm10 on vm10 2026-03-09T17:27:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:40 vm11 ceph-mon[53973]: mgrmap e18: vm10.qompzp(active, since 11s), standbys: vm11.ohhvcs 2026-03-09T17:27:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr metadata", "who": "vm11.ohhvcs", "id": "vm11.ohhvcs"}]: dispatch 2026-03-09T17:27:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:27:40.973 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:27:40.973 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":2,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","modified":"2026-03-09T17:27:34.290410Z","created":"2026-03-09T17:26:13.054583Z","min_mon_release":17,"min_mon_release_name":"quincy","election_strategy":1,"disallowed_leaders: ":"","stretch_mode":false,"tiebreaker_mon":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy"],"optional":[]},"mons":[{"rank":0,"name":"vm10","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"vm11","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:3300","nonce":0},{"type":"v1","addr":"192.168.123.111:6789","nonce":0}]},"addr":"192.168.123.111:6789/0","public_addr":"192.168.123.111:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-09T17:27:40.975 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 2 2026-03-09T17:27:41.034 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-09T17:27:41.034 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph config generate-minimal-conf 2026-03-09T17:27:41.626 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:41 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/2263560999' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:41.627 INFO:teuthology.orchestra.run.vm10.stdout:# minimal ceph.conf for 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:27:41.627 INFO:teuthology.orchestra.run.vm10.stdout:[global] 2026-03-09T17:27:41.627 INFO:teuthology.orchestra.run.vm10.stdout: fsid = 01c7db6a-1bdd-11f1-ac39-b134231210ea 2026-03-09T17:27:41.627 INFO:teuthology.orchestra.run.vm10.stdout: mon_host = [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] 2026-03-09T17:27:41.681 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-09T17:27:41.681 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:27:41.681 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T17:27:41.720 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:27:41.720 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:27:41.820 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T17:27:41.820 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T17:27:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:41 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/2263560999' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T17:27:41.852 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T17:27:41.852 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:27:41.920 INFO:tasks.cephadm:Deploying OSDs... 2026-03-09T17:27:41.920 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:27:41.921 DEBUG:teuthology.orchestra.run.vm10:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T17:27:41.935 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:27:41.936 DEBUG:teuthology.orchestra.run.vm10:> ls /dev/[sv]d? 2026-03-09T17:27:41.995 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vda 2026-03-09T17:27:41.995 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdb 2026-03-09T17:27:41.995 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdc 2026-03-09T17:27:41.995 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdd 2026-03-09T17:27:41.995 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vde 2026-03-09T17:27:41.995 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T17:27:41.995 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T17:27:41.995 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdb 2026-03-09T17:27:42.054 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdb 2026-03-09T17:27:42.054 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:27:42.054 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T17:27:42.054 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:27:42.055 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:27:42.055 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 17:26:49.417317995 +0000 2026-03-09T17:27:42.055 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 17:26:48.993317533 +0000 2026-03-09T17:27:42.055 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 17:26:48.993317533 +0000 2026-03-09T17:27:42.055 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 17:20:32.274000000 +0000 2026-03-09T17:27:42.055 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T17:27:42.122 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T17:27:42.122 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T17:27:42.122 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000172262 s, 3.0 MB/s 2026-03-09T17:27:42.123 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T17:27:42.182 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdc 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdc 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 17:26:49.512318099 +0000 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 17:26:49.003317544 +0000 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 17:26:49.003317544 +0000 2026-03-09T17:27:42.244 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 17:20:32.284000000 +0000 2026-03-09T17:27:42.244 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T17:27:42.315 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T17:27:42.315 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T17:27:42.315 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000136385 s, 3.8 MB/s 2026-03-09T17:27:42.316 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T17:27:42.378 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdd 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdd 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 17:26:49.619318215 +0000 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 17:26:49.007317548 +0000 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 17:26:49.007317548 +0000 2026-03-09T17:27:42.439 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 17:20:32.289000000 +0000 2026-03-09T17:27:42.439 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T17:27:42.507 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T17:27:42.507 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T17:27:42.507 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000145923 s, 3.5 MB/s 2026-03-09T17:27:42.508 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T17:27:42.568 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vde 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vde 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 17:26:49.724318330 +0000 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 17:26:49.012317554 +0000 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 17:26:49.012317554 +0000 2026-03-09T17:27:42.627 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 17:20:32.355000000 +0000 2026-03-09T17:27:42.627 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T17:27:42.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:42 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/2775160211' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:42.694 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T17:27:42.694 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T17:27:42.694 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000216958 s, 2.4 MB/s 2026-03-09T17:27:42.695 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T17:27:42.756 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T17:27:42.756 DEBUG:teuthology.orchestra.run.vm11:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T17:27:42.775 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T17:27:42.775 DEBUG:teuthology.orchestra.run.vm11:> ls /dev/[sv]d? 2026-03-09T17:27:42.831 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vda 2026-03-09T17:27:42.831 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vdb 2026-03-09T17:27:42.831 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vdc 2026-03-09T17:27:42.831 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vdd 2026-03-09T17:27:42.831 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vde 2026-03-09T17:27:42.831 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T17:27:42.831 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T17:27:42.832 DEBUG:teuthology.orchestra.run.vm11:> stat /dev/vdb 2026-03-09T17:27:42.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:42 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/2775160211' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout: File: /dev/vdb 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout:Access: 2026-03-09 17:27:30.764851777 +0000 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout:Modify: 2026-03-09 17:27:30.437851267 +0000 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout:Change: 2026-03-09 17:27:30.437851267 +0000 2026-03-09T17:27:42.856 INFO:teuthology.orchestra.run.vm11.stdout: Birth: 2026-03-09 17:21:56.240000000 +0000 2026-03-09T17:27:42.856 DEBUG:teuthology.orchestra.run.vm11:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T17:27:42.921 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records in 2026-03-09T17:27:42.921 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records out 2026-03-09T17:27:42.921 INFO:teuthology.orchestra.run.vm11.stderr:512 bytes copied, 0.000184886 s, 2.8 MB/s 2026-03-09T17:27:42.922 DEBUG:teuthology.orchestra.run.vm11:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T17:27:42.979 DEBUG:teuthology.orchestra.run.vm11:> stat /dev/vdc 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout: File: /dev/vdc 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout:Access: 2026-03-09 17:27:30.853851916 +0000 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout:Modify: 2026-03-09 17:27:30.446851281 +0000 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout:Change: 2026-03-09 17:27:30.446851281 +0000 2026-03-09T17:27:43.036 INFO:teuthology.orchestra.run.vm11.stdout: Birth: 2026-03-09 17:21:56.249000000 +0000 2026-03-09T17:27:43.036 DEBUG:teuthology.orchestra.run.vm11:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T17:27:43.100 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records in 2026-03-09T17:27:43.100 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records out 2026-03-09T17:27:43.100 INFO:teuthology.orchestra.run.vm11.stderr:512 bytes copied, 0.000201937 s, 2.5 MB/s 2026-03-09T17:27:43.101 DEBUG:teuthology.orchestra.run.vm11:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T17:27:43.158 DEBUG:teuthology.orchestra.run.vm11:> stat /dev/vdd 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout: File: /dev/vdd 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout:Access: 2026-03-09 17:27:30.911852007 +0000 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout:Modify: 2026-03-09 17:27:30.443851276 +0000 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout:Change: 2026-03-09 17:27:30.443851276 +0000 2026-03-09T17:27:43.217 INFO:teuthology.orchestra.run.vm11.stdout: Birth: 2026-03-09 17:21:56.256000000 +0000 2026-03-09T17:27:43.217 DEBUG:teuthology.orchestra.run.vm11:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T17:27:43.281 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records in 2026-03-09T17:27:43.281 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records out 2026-03-09T17:27:43.281 INFO:teuthology.orchestra.run.vm11.stderr:512 bytes copied, 0.000193241 s, 2.6 MB/s 2026-03-09T17:27:43.282 DEBUG:teuthology.orchestra.run.vm11:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T17:27:43.339 DEBUG:teuthology.orchestra.run.vm11:> stat /dev/vde 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout: File: /dev/vde 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout:Access: 2026-03-09 17:27:30.982852118 +0000 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout:Modify: 2026-03-09 17:27:30.442851274 +0000 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout:Change: 2026-03-09 17:27:30.442851274 +0000 2026-03-09T17:27:43.396 INFO:teuthology.orchestra.run.vm11.stdout: Birth: 2026-03-09 17:21:56.297000000 +0000 2026-03-09T17:27:43.396 DEBUG:teuthology.orchestra.run.vm11:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T17:27:43.461 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records in 2026-03-09T17:27:43.461 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records out 2026-03-09T17:27:43.461 INFO:teuthology.orchestra.run.vm11.stderr:512 bytes copied, 0.000196819 s, 2.6 MB/s 2026-03-09T17:27:43.463 DEBUG:teuthology.orchestra.run.vm11:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T17:27:43.520 INFO:tasks.cephadm:Deploying osd.0 on vm10 with /dev/vde... 2026-03-09T17:27:43.520 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- lvm zap /dev/vde 2026-03-09T17:27:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:49 vm11 ceph-mon[53973]: pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:49.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:49 vm10 ceph-mon[53712]: pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:51 vm10 ceph-mon[53712]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:51 vm11 ceph-mon[53973]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:53 vm11 ceph-mon[53973]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:53 vm10 ceph-mon[53712]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:55 vm11 ceph-mon[53973]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:55.366 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:55 vm10 ceph-mon[53712]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:56.292 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:27:56.309 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch daemon add osd vm10:/dev/vde 2026-03-09T17:27:57.032 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:57.032 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:27:57.032 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:57.032 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:27:57.032 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:56 vm10 ceph-mon[53712]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:57.032 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:27:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:27:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:56 vm11 ceph-mon[53973]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:57.818 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: Updating vm11:/etc/ceph/ceph.conf 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: Updating vm11:/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='client.14264 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:27:57.819 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: Updating vm11:/etc/ceph/ceph.conf 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: Updating vm11:/etc/ceph/ceph.client.admin.keyring 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='client.14264 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: Reconfiguring mon.vm10 (unknown last config time)... 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: Reconfiguring daemon mon.vm10 on vm10 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/1938231017' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ac80623-9e73-4d06-ab1e-44875a280f2c"}]: dispatch 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/1938231017' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6ac80623-9e73-4d06-ab1e-44875a280f2c"}]': finished 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:27:58 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/1573259221' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:27:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T17:27:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: Reconfiguring mon.vm10 (unknown last config time)... 2026-03-09T17:27:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: Reconfiguring daemon mon.vm10 on vm10 2026-03-09T17:27:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/1938231017' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ac80623-9e73-4d06-ab1e-44875a280f2c"}]: dispatch 2026-03-09T17:27:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/1938231017' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6ac80623-9e73-4d06-ab1e-44875a280f2c"}]': finished 2026-03-09T17:27:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T17:27:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:27:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:27:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:27:58 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/1573259221' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:00.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:00 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:00.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:00 vm10 ceph-mon[53712]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:00.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:00 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:00.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:00 vm11 ceph-mon[53973]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: Reconfiguring mgr.vm10.qompzp (unknown last config time)... 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm10.qompzp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: Reconfiguring daemon mgr.vm10.qompzp on vm10 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T17:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: Reconfiguring mgr.vm10.qompzp (unknown last config time)... 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm10.qompzp", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: Reconfiguring daemon mgr.vm10.qompzp on vm10 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T17:28:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: Reconfiguring alertmanager.vm10 (dependencies changed)... 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: Reconfiguring daemon alertmanager.vm10 on vm10 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: Deploying daemon osd.0 on vm10 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: Reconfiguring crash.vm10 (monmap changed)... 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm10", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:03 vm11 ceph-mon[53973]: Reconfiguring daemon crash.vm10 on vm10 2026-03-09T17:28:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: Reconfiguring alertmanager.vm10 (dependencies changed)... 2026-03-09T17:28:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: Reconfiguring daemon alertmanager.vm10 on vm10 2026-03-09T17:28:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: Deploying daemon osd.0 on vm10 2026-03-09T17:28:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: Reconfiguring crash.vm10 (monmap changed)... 2026-03-09T17:28:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm10", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T17:28:03.686 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:03.686 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:03 vm10 ceph-mon[53712]: Reconfiguring daemon crash.vm10 on vm10 2026-03-09T17:28:05.053 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 0 on host 'vm10' 2026-03-09T17:28:05.136 DEBUG:teuthology.orchestra.run.vm10:osd.0> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@osd.0.service 2026-03-09T17:28:05.139 INFO:tasks.cephadm:Deploying osd.1 on vm10 with /dev/vdd... 2026-03-09T17:28:05.139 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- lvm zap /dev/vdd 2026-03-09T17:28:05.221 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:05.221 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:04 vm10 ceph-mon[53712]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:05.221 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:04 vm11 ceph-mon[53973]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: Reconfiguring grafana.vm10 (dependencies changed)... 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: Reconfiguring daemon grafana.vm10 on vm10 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm11", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm11.ohhvcs", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:28:06.053 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:06.053 INFO:journalctl@ceph.osd.0.vm10.stdout:Mar 09 17:28:05 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-0[63976]: 2026-03-09T17:28:05.849+0000 7f2181c963c0 -1 osd.0 0 log_to_monitors true 2026-03-09T17:28:06.066 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: Reconfiguring grafana.vm10 (dependencies changed)... 2026-03-09T17:28:06.066 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: Reconfiguring daemon grafana.vm10 on vm10 2026-03-09T17:28:06.066 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:06.066 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:06.066 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm11", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-09T17:28:06.066 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:06.066 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T17:28:06.067 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:06.067 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm11.ohhvcs", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T17:28:06.067 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T17:28:06.067 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:06.136 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:28:06.152 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch daemon add osd vm10:/dev/vdd 2026-03-09T17:28:07.551 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: Reconfiguring crash.vm11 (monmap changed)... 2026-03-09T17:28:07.551 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: Reconfiguring daemon crash.vm11 on vm11 2026-03-09T17:28:07.551 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: Reconfiguring mgr.vm11.ohhvcs (monmap changed)... 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: Reconfiguring daemon mgr.vm11.ohhvcs on vm11 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: Reconfiguring mon.vm11 (monmap changed)... 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: Reconfiguring daemon mon.vm11 on vm11 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.110:9093"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.110:3000"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.110:9095"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:07.552 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: Reconfiguring crash.vm11 (monmap changed)... 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: Reconfiguring daemon crash.vm11 on vm11 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: Reconfiguring mgr.vm11.ohhvcs (monmap changed)... 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: Reconfiguring daemon mgr.vm11.ohhvcs on vm11 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: Reconfiguring mon.vm11 (monmap changed)... 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: Reconfiguring daemon mon.vm11 on vm11 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.110:9093"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.110:3000"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.110:9095"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:07.928 INFO:journalctl@ceph.osd.0.vm10.stdout:Mar 09 17:28:07 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-0[63976]: 2026-03-09T17:28:07.664+0000 7f2178699700 -1 osd.0 0 waiting for initial osdmap 2026-03-09T17:28:07.928 INFO:journalctl@ceph.osd.0.vm10.stdout:Mar 09 17:28:07 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-0[63976]: 2026-03-09T17:28:07.671+0000 7f2173030700 -1 osd.0 8 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:28:08.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.110:9093"}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.110:3000"}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.110:9095"}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='client.24109 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/2348619235' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "fc09d5ce-67c5-4201-a2f2-fe1cb3b21fad"}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/2348619235' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "fc09d5ce-67c5-4201-a2f2-fe1cb3b21fad"}]': finished 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: osdmap e8: 2 total, 0 up, 2 in 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:08.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:08 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/596773260' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://192.168.123.110:9093"}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://192.168.123.110:3000"}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://192.168.123.110:9095"}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='client.24109 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/2348619235' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "fc09d5ce-67c5-4201-a2f2-fe1cb3b21fad"}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/2348619235' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "fc09d5ce-67c5-4201-a2f2-fe1cb3b21fad"}]': finished 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: osdmap e8: 2 total, 0 up, 2 in 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:08.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:08 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/596773260' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:09 vm10 ceph-mon[53712]: purged_snaps scrub starts 2026-03-09T17:28:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:09 vm10 ceph-mon[53712]: purged_snaps scrub ok 2026-03-09T17:28:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:09 vm10 ceph-mon[53712]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:09 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:09 vm10 ceph-mon[53712]: osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049] boot 2026-03-09T17:28:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:09 vm10 ceph-mon[53712]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T17:28:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:09 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:09 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:09 vm11 ceph-mon[53973]: purged_snaps scrub starts 2026-03-09T17:28:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:09 vm11 ceph-mon[53973]: purged_snaps scrub ok 2026-03-09T17:28:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:09 vm11 ceph-mon[53973]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T17:28:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:09 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:09 vm11 ceph-mon[53973]: osd.0 [v2:192.168.123.110:6802/1156255049,v1:192.168.123.110:6803/1156255049] boot 2026-03-09T17:28:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:09 vm11 ceph-mon[53973]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T17:28:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:09 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T17:28:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:09 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:10 vm11 ceph-mon[53973]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T17:28:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:10 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:10 vm11 ceph-mon[53973]: pgmap v19: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:11.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:10 vm10 ceph-mon[53712]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T17:28:11.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:10 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:11.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:10 vm10 ceph-mon[53712]: pgmap v19: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: Detected new or changed devices on vm10 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: Deploying daemon osd.1 on vm10 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: pgmap v20: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:12.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:12.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: Detected new or changed devices on vm10 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: Deploying daemon osd.1 on vm10 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: pgmap v20: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:14.038 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:14 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:14.038 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:14 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:14.038 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:14 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:14.038 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:14 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:14.038 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:14 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:14.038 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:14 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:14.418 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:14 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:14.418 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:14 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:14.418 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:14 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:14.419 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:14 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:14.419 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:14 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:14.419 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:14 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:14.614 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 1 on host 'vm10' 2026-03-09T17:28:14.684 DEBUG:teuthology.orchestra.run.vm10:osd.1> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@osd.1.service 2026-03-09T17:28:14.686 INFO:tasks.cephadm:Deploying osd.2 on vm10 with /dev/vdc... 2026-03-09T17:28:14.686 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- lvm zap /dev/vdc 2026-03-09T17:28:15.387 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 09 17:28:15 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-1[67500]: 2026-03-09T17:28:15.332+0000 7fd37efdb3c0 -1 osd.1 0 log_to_monitors true 2026-03-09T17:28:15.387 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:15 vm10 ceph-mon[53712]: pgmap v21: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:15.387 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:15 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:15.387 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:15 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:15 vm11 ceph-mon[53973]: pgmap v21: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:15 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:15 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:15.765 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:28:15.784 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch daemon add osd vm10:/dev/vdc 2026-03-09T17:28:16.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:16 vm11 ceph-mon[53973]: from='osd.1 [v2:192.168.123.110:6810/1744031210,v1:192.168.123.110:6811/1744031210]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T17:28:16.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:16 vm11 ceph-mon[53973]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T17:28:16.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:16.611 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:16 vm10 ceph-mon[53712]: from='osd.1 [v2:192.168.123.110:6810/1744031210,v1:192.168.123.110:6811/1744031210]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T17:28:16.611 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:16 vm10 ceph-mon[53712]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T17:28:16.611 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: pgmap v22: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='osd.1 [v2:192.168.123.110:6810/1744031210,v1:192.168.123.110:6811/1744031210]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:17.417 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 09 17:28:17 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-1[67500]: 2026-03-09T17:28:17.314+0000 7fd3771e1700 -1 osd.1 0 waiting for initial osdmap 2026-03-09T17:28:17.417 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 09 17:28:17 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-1[67500]: 2026-03-09T17:28:17.321+0000 7fd371b78700 -1 osd.1 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:28:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: pgmap v22: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='osd.1 [v2:192.168.123.110:6810/1744031210,v1:192.168.123.110:6811/1744031210]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='client.14296 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: osdmap e12: 2 total, 1 up, 2 in 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/372956316' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9de17247-790c-41ec-9cd7-a62e65ec8735"}]: dispatch 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9de17247-790c-41ec-9cd7-a62e65ec8735"}]: dispatch 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: osd.1 [v2:192.168.123.110:6810/1744031210,v1:192.168.123.110:6811/1744031210] boot 2026-03-09T17:28:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9de17247-790c-41ec-9cd7-a62e65ec8735"}]': finished 2026-03-09T17:28:18.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T17:28:18.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:18.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:18.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/748444897' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:18.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:18.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:18 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='client.14296 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: osdmap e12: 2 total, 1 up, 2 in 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/372956316' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9de17247-790c-41ec-9cd7-a62e65ec8735"}]: dispatch 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "9de17247-790c-41ec-9cd7-a62e65ec8735"}]: dispatch 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: osd.1 [v2:192.168.123.110:6810/1744031210,v1:192.168.123.110:6811/1744031210] boot 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "9de17247-790c-41ec-9cd7-a62e65ec8735"}]': finished 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/748444897' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:18.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:18 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:19 vm11 ceph-mon[53973]: purged_snaps scrub starts 2026-03-09T17:28:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:19 vm11 ceph-mon[53973]: purged_snaps scrub ok 2026-03-09T17:28:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:19 vm11 ceph-mon[53973]: pgmap v26: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:19 vm11 ceph-mon[53973]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T17:28:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:19 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:19 vm10 ceph-mon[53712]: purged_snaps scrub starts 2026-03-09T17:28:19.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:19 vm10 ceph-mon[53712]: purged_snaps scrub ok 2026-03-09T17:28:19.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:19 vm10 ceph-mon[53712]: pgmap v26: 0 pgs: ; 0 B data, 4.8 MiB used, 20 GiB / 20 GiB avail 2026-03-09T17:28:19.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:19 vm10 ceph-mon[53712]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T17:28:19.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:19 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:21 vm10 ceph-mon[53712]: pgmap v28: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:28:21.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:21 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T17:28:21.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:21 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:21 vm11 ceph-mon[53973]: pgmap v28: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:28:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:21 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T17:28:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:21 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:22.578 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:22 vm10 ceph-mon[53712]: Deploying daemon osd.2 on vm10 2026-03-09T17:28:22.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:22 vm11 ceph-mon[53973]: Deploying daemon osd.2 on vm10 2026-03-09T17:28:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:23 vm10 ceph-mon[53712]: pgmap v29: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:28:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:23 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:23 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:23 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:23 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:23.565 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 2 on host 'vm10' 2026-03-09T17:28:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:23 vm11 ceph-mon[53973]: pgmap v29: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:28:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:23 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:23 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:23 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:23 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:23.665 DEBUG:teuthology.orchestra.run.vm10:osd.2> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@osd.2.service 2026-03-09T17:28:23.674 INFO:tasks.cephadm:Deploying osd.3 on vm10 with /dev/vdb... 2026-03-09T17:28:23.674 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- lvm zap /dev/vdb 2026-03-09T17:28:24.632 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:28:24.649 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch daemon add osd vm10:/dev/vdb 2026-03-09T17:28:24.818 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:24 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:24.818 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:24 vm10 ceph-mon[53712]: pgmap v30: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:28:24.818 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:24 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:24.819 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 17:28:24 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-2[70809]: 2026-03-09T17:28:24.755+0000 7f7a63ce23c0 -1 osd.2 0 log_to_monitors true 2026-03-09T17:28:24.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:24 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:24.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:24 vm11 ceph-mon[53973]: pgmap v30: 0 pgs: ; 0 B data, 9.7 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:28:24.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:24 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:25.616 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:25 vm10 ceph-mon[53712]: from='osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T17:28:25.616 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:25 vm10 ceph-mon[53712]: from='client.14314 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:25.616 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:25 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:25.616 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:25 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:25.616 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:25 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:25 vm11 ceph-mon[53973]: from='osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T17:28:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:25 vm11 ceph-mon[53973]: from='client.14314 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:25 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:25 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:25 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: Detected new or changed devices on vm10 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: pgmap v32: 0 pgs: ; 0 B data, 9.8 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/2028649781' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5b05874b-c02b-475c-8aa7-e90fa7ea69b7"}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5b05874b-c02b-475c-8aa7-e90fa7ea69b7"}]: dispatch 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T17:28:26.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5b05874b-c02b-475c-8aa7-e90fa7ea69b7"}]': finished 2026-03-09T17:28:26.584 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 17:28:26 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-2[70809]: 2026-03-09T17:28:26.300+0000 7f7a5a6e5700 -1 osd.2 0 waiting for initial osdmap 2026-03-09T17:28:26.584 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 17:28:26 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-2[70809]: 2026-03-09T17:28:26.306+0000 7f7a5607e700 -1 osd.2 16 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:28:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T17:28:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T17:28:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: Detected new or changed devices on vm10 2026-03-09T17:28:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: pgmap v32: 0 pgs: ; 0 B data, 9.8 MiB used, 40 GiB / 40 GiB avail 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/2028649781' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5b05874b-c02b-475c-8aa7-e90fa7ea69b7"}]: dispatch 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5b05874b-c02b-475c-8aa7-e90fa7ea69b7"}]: dispatch 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5b05874b-c02b-475c-8aa7-e90fa7ea69b7"}]': finished 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: osdmap e16: 4 total, 2 up, 4 in 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:26.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:26 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: osdmap e16: 4 total, 2 up, 4 in 2026-03-09T17:28:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:26 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:27 vm11 ceph-mon[53973]: purged_snaps scrub starts 2026-03-09T17:28:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:27 vm11 ceph-mon[53973]: purged_snaps scrub ok 2026-03-09T17:28:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:27 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/1230092333' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:27 vm11 ceph-mon[53973]: osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350] boot 2026-03-09T17:28:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:27 vm11 ceph-mon[53973]: osdmap e17: 4 total, 3 up, 4 in 2026-03-09T17:28:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:27 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:27 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:27 vm10 ceph-mon[53712]: purged_snaps scrub starts 2026-03-09T17:28:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:27 vm10 ceph-mon[53712]: purged_snaps scrub ok 2026-03-09T17:28:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:27 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/1230092333' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:27 vm10 ceph-mon[53712]: osd.2 [v2:192.168.123.110:6818/3248441350,v1:192.168.123.110:6819/3248441350] boot 2026-03-09T17:28:28.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:27 vm10 ceph-mon[53712]: osdmap e17: 4 total, 3 up, 4 in 2026-03-09T17:28:28.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:27 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T17:28:28.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:27 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:29.034 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:28 vm10 ceph-mon[53712]: pgmap v35: 0 pgs: ; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:29.034 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]: dispatch 2026-03-09T17:28:29.034 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:28:29.034 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:28:29.034 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]': finished 2026-03-09T17:28:29.034 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:28 vm10 ceph-mon[53712]: osdmap e18: 4 total, 3 up, 4 in 2026-03-09T17:28:29.034 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:29.034 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T17:28:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:28 vm11 ceph-mon[53973]: pgmap v35: 0 pgs: ; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]: dispatch 2026-03-09T17:28:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:28:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:28:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32}]': finished 2026-03-09T17:28:29.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:28 vm11 ceph-mon[53973]: osdmap e18: 4 total, 3 up, 4 in 2026-03-09T17:28:29.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:29.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T17:28:30.428 INFO:journalctl@ceph.osd.0.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73514]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vde 2026-03-09T17:28:30.428 INFO:journalctl@ceph.osd.0.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73514]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T17:28:30.428 INFO:journalctl@ceph.osd.0.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73514]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:28:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T17:28:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 ceph-mon[53712]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T17:28:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T17:28:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 ceph-mon[53712]: Deploying daemon osd.3 on vm10 2026-03-09T17:28:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 ceph-mon[53712]: pgmap v38: 1 pgs: 1 unknown; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:30.685 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73545]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vdd 2026-03-09T17:28:30.685 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73545]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T17:28:30.685 INFO:journalctl@ceph.osd.1.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73545]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:28:30.685 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73602]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vdc 2026-03-09T17:28:30.685 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73602]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T17:28:30.685 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73602]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:28:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:30 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T17:28:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:30 vm11 ceph-mon[53973]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T17:28:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:30 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:30 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T17:28:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:30 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:30.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:30 vm11 ceph-mon[53973]: Deploying daemon osd.3 on vm10 2026-03-09T17:28:30.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:30 vm11 ceph-mon[53973]: pgmap v38: 1 pgs: 1 unknown; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:31.075 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73646]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vda 2026-03-09T17:28:31.075 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73646]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T17:28:31.075 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:30 vm10 sudo[73646]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:28:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 sudo[56187]: ceph : TTY=unknown ; PWD=/ ; USER=root ; COMMAND=/sbin/smartctl -x --json=o /dev/vda 2026-03-09T17:28:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 sudo[56187]: pam_unix(sudo:session): session opened for user root by (uid=0) 2026-03-09T17:28:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 sudo[56187]: pam_unix(sudo:session): session closed for user root 2026-03-09T17:28:31.750 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:28:31.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:31 vm10 ceph-mon[53712]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:28:31.831 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T17:28:31.831 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:31.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:28:31.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:28:31.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T17:28:31.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:28:31.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T17:28:31.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm10"}]: dispatch 2026-03-09T17:28:31.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mon metadata", "id": "vm11"}]: dispatch 2026-03-09T17:28:31.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:31 vm11 ceph-mon[53973]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T17:28:32.915 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 3 on host 'vm10' 2026-03-09T17:28:32.982 DEBUG:teuthology.orchestra.run.vm10:osd.3> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@osd.3.service 2026-03-09T17:28:32.991 INFO:tasks.cephadm:Deploying osd.4 on vm11 with /dev/vde... 2026-03-09T17:28:32.991 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- lvm zap /dev/vde 2026-03-09T17:28:33.026 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:33.026 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:33.026 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:33.026 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:32 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:33.026 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:32 vm10 ceph-mon[53712]: pgmap v40: 1 pgs: 1 unknown; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:33.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:32 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:33.099 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:32 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:33.099 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:32 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:33.099 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:32 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:33.099 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:32 vm11 ceph-mon[53973]: pgmap v40: 1 pgs: 1 unknown; 0 B data, 15 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:33.877 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:33 vm10 ceph-mon[53712]: mgrmap e19: vm10.qompzp(active, since 65s), standbys: vm11.ohhvcs 2026-03-09T17:28:33.877 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:33 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:33 vm11 ceph-mon[53973]: mgrmap e19: vm10.qompzp(active, since 65s), standbys: vm11.ohhvcs 2026-03-09T17:28:34.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:33 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:34.185 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 17:28:33 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-3[73804]: 2026-03-09T17:28:33.893+0000 7fdb939413c0 -1 osd.3 0 log_to_monitors true 2026-03-09T17:28:34.277 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:28:34.292 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch daemon add osd vm11:/dev/vde 2026-03-09T17:28:35.412 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:35 vm11 ceph-mon[53973]: from='osd.3 [v2:192.168.123.110:6826/2088446903,v1:192.168.123.110:6827/2088446903]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T17:28:35.412 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:35 vm11 ceph-mon[53973]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T17:28:35.412 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:35 vm11 ceph-mon[53973]: pgmap v41: 1 pgs: 1 active+clean; 449 KiB data, 17 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:35.412 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:35 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:35.412 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:35 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:35.412 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:35 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:35.412 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:35 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:35 vm10 ceph-mon[53712]: from='osd.3 [v2:192.168.123.110:6826/2088446903,v1:192.168.123.110:6827/2088446903]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T17:28:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:35 vm10 ceph-mon[53712]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T17:28:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:35 vm10 ceph-mon[53712]: pgmap v41: 1 pgs: 1 active+clean; 449 KiB data, 17 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:35 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:35 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:35 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:35 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:36.185 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 17:28:35 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-3[73804]: 2026-03-09T17:28:35.953+0000 7fdb8a344700 -1 osd.3 0 waiting for initial osdmap 2026-03-09T17:28:36.185 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 17:28:35 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-3[73804]: 2026-03-09T17:28:35.960+0000 7fdb844da700 -1 osd.3 22 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='client.24145 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='osd.3 [v2:192.168.123.110:6826/2088446903,v1:192.168.123.110:6827/2088446903]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/3687975143' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa"}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa"}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa"}]': finished 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: osdmap e22: 5 total, 3 up, 5 in 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:36.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:36 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='client.24145 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='osd.3 [v2:192.168.123.110:6826/2088446903,v1:192.168.123.110:6827/2088446903]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/3687975143' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa"}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa"}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa"}]': finished 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: osdmap e22: 5 total, 3 up, 5 in 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:36.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:36 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: purged_snaps scrub starts 2026-03-09T17:28:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: purged_snaps scrub ok 2026-03-09T17:28:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: Detected new or changed devices on vm10 2026-03-09T17:28:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: pgmap v44: 1 pgs: 1 active+clean; 449 KiB data, 17 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/242823826' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:37.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: osd.3 [v2:192.168.123.110:6826/2088446903,v1:192.168.123.110:6827/2088446903] boot 2026-03-09T17:28:37.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: osdmap e23: 5 total, 4 up, 5 in 2026-03-09T17:28:37.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:37.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:37 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: purged_snaps scrub starts 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: purged_snaps scrub ok 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: Detected new or changed devices on vm10 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: pgmap v44: 1 pgs: 1 active+clean; 449 KiB data, 17 MiB used, 60 GiB / 60 GiB avail 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/242823826' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: osd.3 [v2:192.168.123.110:6826/2088446903,v1:192.168.123.110:6827/2088446903] boot 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: osdmap e23: 5 total, 4 up, 5 in 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T17:28:37.943 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:37 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:38.719 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:38 vm11 ceph-mon[53973]: pgmap v46: 1 pgs: 1 active+clean; 449 KiB data, 22 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:38 vm10 ceph-mon[53712]: pgmap v46: 1 pgs: 1 active+clean; 449 KiB data, 22 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:39.719 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:39 vm11 ceph-mon[53973]: osdmap e24: 5 total, 4 up, 5 in 2026-03-09T17:28:39.719 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:39 vm10 ceph-mon[53712]: osdmap e24: 5 total, 4 up, 5 in 2026-03-09T17:28:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:40.528 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T17:28:40.529 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:40.529 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:40 vm11 ceph-mon[53973]: Deploying daemon osd.4 on vm11 2026-03-09T17:28:40.529 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:40 vm11 ceph-mon[53973]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T17:28:40.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:40.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:40 vm10 ceph-mon[53712]: Deploying daemon osd.4 on vm11 2026-03-09T17:28:40.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:40 vm10 ceph-mon[53712]: pgmap v48: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:42.293 INFO:teuthology.orchestra.run.vm11.stdout:Created osd(s) 4 on host 'vm11' 2026-03-09T17:28:42.354 DEBUG:teuthology.orchestra.run.vm11:osd.4> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@osd.4.service 2026-03-09T17:28:42.356 INFO:tasks.cephadm:Deploying osd.5 on vm11 with /dev/vdd... 2026-03-09T17:28:42.356 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- lvm zap /dev/vdd 2026-03-09T17:28:42.701 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:42 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:42.701 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:42 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:42.701 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:42 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:42.701 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:42 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:42.701 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:42 vm11 ceph-mon[53973]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:42.701 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:42 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:42.701 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:42 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:42.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:42 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:42.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:42 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:42.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:42 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:42.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:42 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:42.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:42 vm10 ceph-mon[53712]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:42.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:42 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:42.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:42 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:43.109 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:28:43.123 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch daemon add osd vm11:/dev/vdd 2026-03-09T17:28:43.910 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 17:28:43 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-4[58005]: 2026-03-09T17:28:43.782+0000 7fae489d63c0 -1 osd.4 0 log_to_monitors true 2026-03-09T17:28:44.183 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:43 vm11 ceph-mon[53973]: from='osd.4 [v2:192.168.123.111:6800/236530777,v1:192.168.123.111:6801/236530777]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T17:28:44.183 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:43 vm11 ceph-mon[53973]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T17:28:44.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:43 vm10 ceph-mon[53712]: from='osd.4 [v2:192.168.123.111:6800/236530777,v1:192.168.123.111:6801/236530777]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T17:28:44.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:43 vm10 ceph-mon[53712]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T17:28:45.233 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 17:28:45 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-4[58005]: 2026-03-09T17:28:45.156+0000 7fae3f3d9700 -1 osd.4 0 waiting for initial osdmap 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: osdmap e25: 5 total, 4 up, 5 in 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='osd.4 [v2:192.168.123.111:6800/236530777,v1:192.168.123.111:6801/236530777]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='client.24163 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: Detected new or changed devices on vm11 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:45.591 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 17:28:45 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-4[58005]: 2026-03-09T17:28:45.237+0000 7fae3bd74700 -1 osd.4 26 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:28:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T17:28:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: osdmap e25: 5 total, 4 up, 5 in 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='osd.4 [v2:192.168.123.111:6800/236530777,v1:192.168.123.111:6801/236530777]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='client.24163 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: Detected new or changed devices on vm11 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/1270244026' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "015dd696-8386-4fa0-b5f2-f22dc317f27f"}]: dispatch 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "015dd696-8386-4fa0-b5f2-f22dc317f27f"}]: dispatch 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: osd.4 [v2:192.168.123.111:6800/236530777,v1:192.168.123.111:6801/236530777] boot 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "015dd696-8386-4fa0-b5f2-f22dc317f27f"}]': finished 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: osdmap e27: 6 total, 5 up, 6 in 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:46 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/4235329750' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T17:28:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/1270244026' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "015dd696-8386-4fa0-b5f2-f22dc317f27f"}]: dispatch 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "015dd696-8386-4fa0-b5f2-f22dc317f27f"}]: dispatch 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: osd.4 [v2:192.168.123.111:6800/236530777,v1:192.168.123.111:6801/236530777] boot 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "015dd696-8386-4fa0-b5f2-f22dc317f27f"}]': finished 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: osdmap e27: 6 total, 5 up, 6 in 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:46.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:46 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/4235329750' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:47 vm11 ceph-mon[53973]: purged_snaps scrub starts 2026-03-09T17:28:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:47 vm11 ceph-mon[53973]: purged_snaps scrub ok 2026-03-09T17:28:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:47 vm11 ceph-mon[53973]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:47 vm11 ceph-mon[53973]: osdmap e28: 6 total, 5 up, 6 in 2026-03-09T17:28:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:47 vm10 ceph-mon[53712]: purged_snaps scrub starts 2026-03-09T17:28:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:47 vm10 ceph-mon[53712]: purged_snaps scrub ok 2026-03-09T17:28:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:47 vm10 ceph-mon[53712]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 23 MiB used, 80 GiB / 80 GiB avail 2026-03-09T17:28:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:47 vm10 ceph-mon[53712]: osdmap e28: 6 total, 5 up, 6 in 2026-03-09T17:28:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:48.737 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:48 vm11 ceph-mon[53973]: osdmap e29: 6 total, 5 up, 6 in 2026-03-09T17:28:48.737 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:48 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:48.737 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:48 vm11 ceph-mon[53973]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:28:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:48 vm10 ceph-mon[53712]: osdmap e29: 6 total, 5 up, 6 in 2026-03-09T17:28:49.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:48 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:49.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:48 vm10 ceph-mon[53712]: pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:28:50.308 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:50 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T17:28:50.309 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:50 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:50.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:50 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T17:28:50.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:50 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:51.592 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:51 vm11 ceph-mon[53973]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail; 79 KiB/s, 0 objects/s recovering 2026-03-09T17:28:51.592 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:51 vm11 ceph-mon[53973]: Deploying daemon osd.5 on vm11 2026-03-09T17:28:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:51 vm10 ceph-mon[53712]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail; 79 KiB/s, 0 objects/s recovering 2026-03-09T17:28:51.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:51 vm10 ceph-mon[53712]: Deploying daemon osd.5 on vm11 2026-03-09T17:28:52.958 INFO:teuthology.orchestra.run.vm11.stdout:Created osd(s) 5 on host 'vm11' 2026-03-09T17:28:53.025 DEBUG:teuthology.orchestra.run.vm11:osd.5> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@osd.5.service 2026-03-09T17:28:53.026 INFO:tasks.cephadm:Deploying osd.6 on vm11 with /dev/vdc... 2026-03-09T17:28:53.026 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- lvm zap /dev/vdc 2026-03-09T17:28:53.275 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:53 vm11 ceph-mon[53973]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 29 MiB used, 100 GiB / 100 GiB avail; 62 KiB/s, 0 objects/s recovering 2026-03-09T17:28:53.275 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:53.275 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:53.275 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:53.275 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:53.275 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:53.275 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:53.275 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 17:28:53 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-5[60773]: 2026-03-09T17:28:53.152+0000 7f36a32433c0 -1 osd.5 0 log_to_monitors true 2026-03-09T17:28:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:53 vm10 ceph-mon[53712]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 29 MiB used, 100 GiB / 100 GiB avail; 62 KiB/s, 0 objects/s recovering 2026-03-09T17:28:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:53.965 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:28:53.981 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch daemon add osd vm11:/dev/vdc 2026-03-09T17:28:54.153 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:54 vm11 ceph-mon[53973]: from='osd.5 [v2:192.168.123.111:6808/2564328090,v1:192.168.123.111:6809/2564328090]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:28:54.153 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:54 vm11 ceph-mon[53973]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:28:54.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:54 vm10 ceph-mon[53712]: from='osd.5 [v2:192.168.123.111:6808/2564328090,v1:192.168.123.111:6809/2564328090]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:28:54.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:54 vm10 ceph-mon[53712]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail; 49 KiB/s, 0 objects/s recovering 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: osdmap e30: 6 total, 5 up, 6 in 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='osd.5 [v2:192.168.123.111:6808/2564328090,v1:192.168.123.111:6809/2564328090]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:55.341 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 17:28:55 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-5[60773]: 2026-03-09T17:28:55.071+0000 7f369b449700 -1 osd.5 0 waiting for initial osdmap 2026-03-09T17:28:55.341 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 17:28:55 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-5[60773]: 2026-03-09T17:28:55.088+0000 7f36965e1700 -1 osd.5 31 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:28:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail; 49 KiB/s, 0 objects/s recovering 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: osdmap e30: 6 total, 5 up, 6 in 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='osd.5 [v2:192.168.123.111:6808/2564328090,v1:192.168.123.111:6809/2564328090]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:28:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: from='client.24181 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: Detected new or changed devices on vm11 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/2531600393' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "ac005d54-a0d8-4362-874e-656309d25f84"}]: dispatch 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: osd.5 [v2:192.168.123.111:6808/2564328090,v1:192.168.123.111:6809/2564328090] boot 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/2531600393' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "ac005d54-a0d8-4362-874e-656309d25f84"}]': finished 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: osdmap e32: 7 total, 6 up, 7 in 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:56.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: from='client.24181 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdc", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: Detected new or changed devices on vm11 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/2531600393' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "ac005d54-a0d8-4362-874e-656309d25f84"}]: dispatch 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: osd.5 [v2:192.168.123.111:6808/2564328090,v1:192.168.123.111:6809/2564328090] boot 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/2531600393' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "ac005d54-a0d8-4362-874e-656309d25f84"}]': finished 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: osdmap e32: 7 total, 6 up, 7 in 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T17:28:56.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:28:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:57 vm11 ceph-mon[53973]: purged_snaps scrub starts 2026-03-09T17:28:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:57 vm11 ceph-mon[53973]: purged_snaps scrub ok 2026-03-09T17:28:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:57 vm11 ceph-mon[53973]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:28:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:57 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/179562567' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:57 vm11 ceph-mon[53973]: osdmap e33: 7 total, 6 up, 7 in 2026-03-09T17:28:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:28:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:57 vm10 ceph-mon[53712]: purged_snaps scrub starts 2026-03-09T17:28:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:57 vm10 ceph-mon[53712]: purged_snaps scrub ok 2026-03-09T17:28:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:57 vm10 ceph-mon[53712]: pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 28 MiB used, 100 GiB / 100 GiB avail 2026-03-09T17:28:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:57 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/179562567' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:28:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:57 vm10 ceph-mon[53712]: osdmap e33: 7 total, 6 up, 7 in 2026-03-09T17:28:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:28:58.923 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:58 vm11 ceph-mon[53973]: osdmap e34: 7 total, 6 up, 7 in 2026-03-09T17:28:58.924 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:28:58.924 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:58 vm11 ceph-mon[53973]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:28:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:58 vm10 ceph-mon[53712]: osdmap e34: 7 total, 6 up, 7 in 2026-03-09T17:28:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:28:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:58 vm10 ceph-mon[53712]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:28:59.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:59 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-09T17:28:59.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:59 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:28:59.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:28:59 vm11 ceph-mon[53973]: Deploying daemon osd.6 on vm11 2026-03-09T17:29:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:59 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-09T17:29:00.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:59 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:00.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:28:59 vm10 ceph-mon[53712]: Deploying daemon osd.6 on vm11 2026-03-09T17:29:01.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:00 vm11 ceph-mon[53973]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:29:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:00 vm10 ceph-mon[53712]: pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:29:02.673 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:02 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:02.673 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:02 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:02.673 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:02 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:02.673 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:02 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:02.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:02 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:02.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:02 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:02.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:02 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:02.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:02 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:02.712 INFO:teuthology.orchestra.run.vm11.stdout:Created osd(s) 6 on host 'vm11' 2026-03-09T17:29:02.788 DEBUG:teuthology.orchestra.run.vm11:osd.6> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@osd.6.service 2026-03-09T17:29:02.790 INFO:tasks.cephadm:Deploying osd.7 on vm11 with /dev/vdb... 2026-03-09T17:29:02.790 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- lvm zap /dev/vdb 2026-03-09T17:29:03.684 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:03 vm11 ceph-mon[53973]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:29:03.685 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:03 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:03.685 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:03 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:03 vm10 ceph-mon[53712]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 34 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:29:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:03 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:03 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:04.027 INFO:journalctl@ceph.osd.6.vm11.stdout:Mar 09 17:29:03 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-6[63570]: 2026-03-09T17:29:03.682+0000 7f3d167273c0 -1 osd.6 0 log_to_monitors true 2026-03-09T17:29:04.583 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: from='osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T17:29:04.583 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:29:04.583 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: Detected new or changed devices on vm11 2026-03-09T17:29:04.583 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:04.583 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:29:04.583 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:04.583 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:04.583 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:04.584 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:04.584 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:04 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:04.649 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T17:29:04.664 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch daemon add osd vm11:/dev/vdb 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: from='osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: Detected new or changed devices on vm11 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:04.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:04 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:05.411 INFO:journalctl@ceph.osd.6.vm11.stdout:Mar 09 17:29:05 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-6[63570]: 2026-03-09T17:29:05.367+0000 7f3d0e92d700 -1 osd.6 0 waiting for initial osdmap 2026-03-09T17:29:05.411 INFO:journalctl@ceph.osd.6.vm11.stdout:Mar 09 17:29:05 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-6[63570]: 2026-03-09T17:29:05.385+0000 7f3d072c0700 -1 osd.6 36 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:29:05.411 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:05 vm11 ceph-mon[53973]: from='osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-09T17:29:05.411 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:05 vm11 ceph-mon[53973]: osdmap e35: 7 total, 6 up, 7 in 2026-03-09T17:29:05.411 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:05 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:29:05.411 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:05 vm11 ceph-mon[53973]: from='osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:29:05.411 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:05 vm11 ceph-mon[53973]: from='client.24199 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:05.411 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:05 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:29:05.411 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:05 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:29:05.411 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:05 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:05 vm10 ceph-mon[53712]: from='osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-09T17:29:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:05 vm10 ceph-mon[53712]: osdmap e35: 7 total, 6 up, 7 in 2026-03-09T17:29:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:05 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:29:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:05 vm10 ceph-mon[53712]: from='osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:29:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:05 vm10 ceph-mon[53712]: from='client.24199 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdb", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:05 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T17:29:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:05 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T17:29:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:05 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:06.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: from='osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T17:29:06.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: osdmap e36: 7 total, 6 up, 7 in 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/3021378045' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "eedd693b-c665-4416-af3c-40637ce85fc7"}]: dispatch 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "eedd693b-c665-4416-af3c-40637ce85fc7"}]: dispatch 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969] boot 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "eedd693b-c665-4416-af3c-40637ce85fc7"}]': finished 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: osdmap e37: 8 total, 7 up, 8 in 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:06.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:06 vm10 ceph-mon[53712]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: from='osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: osdmap e36: 7 total, 6 up, 7 in 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/3021378045' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "eedd693b-c665-4416-af3c-40637ce85fc7"}]: dispatch 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "eedd693b-c665-4416-af3c-40637ce85fc7"}]: dispatch 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: osd.6 [v2:192.168.123.111:6816/605495969,v1:192.168.123.111:6817/605495969] boot 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "eedd693b-c665-4416-af3c-40637ce85fc7"}]': finished 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: osdmap e37: 8 total, 7 up, 8 in 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:06.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:06 vm11 ceph-mon[53973]: pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 35 MiB used, 120 GiB / 120 GiB avail 2026-03-09T17:29:07.664 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:07 vm11 ceph-mon[53973]: purged_snaps scrub starts 2026-03-09T17:29:07.664 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:07 vm11 ceph-mon[53973]: purged_snaps scrub ok 2026-03-09T17:29:07.664 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:07 vm11 ceph-mon[53973]: from='client.? 192.168.123.111:0/662695519' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:29:07.664 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:07 vm11 ceph-mon[53973]: osdmap e38: 8 total, 7 up, 8 in 2026-03-09T17:29:07.664 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:07 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:07 vm10 ceph-mon[53712]: purged_snaps scrub starts 2026-03-09T17:29:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:07 vm10 ceph-mon[53712]: purged_snaps scrub ok 2026-03-09T17:29:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:07 vm10 ceph-mon[53712]: from='client.? 192.168.123.111:0/662695519' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T17:29:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:07 vm10 ceph-mon[53712]: osdmap e38: 8 total, 7 up, 8 in 2026-03-09T17:29:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:07 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:08 vm11 ceph-mon[53973]: osdmap e39: 8 total, 7 up, 8 in 2026-03-09T17:29:09.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:08 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:09.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:08 vm11 ceph-mon[53973]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 40 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:29:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:08 vm10 ceph-mon[53712]: osdmap e39: 8 total, 7 up, 8 in 2026-03-09T17:29:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:08 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:08 vm10 ceph-mon[53712]: pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 40 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:29:10.095 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:09 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-09T17:29:10.095 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:09 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:10.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:09 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-09T17:29:10.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:09 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:11.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:10 vm11 ceph-mon[53973]: Deploying daemon osd.7 on vm11 2026-03-09T17:29:11.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:10 vm11 ceph-mon[53973]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 40 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:29:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:10 vm10 ceph-mon[53712]: Deploying daemon osd.7 on vm11 2026-03-09T17:29:11.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:10 vm10 ceph-mon[53712]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 40 MiB used, 140 GiB / 140 GiB avail 2026-03-09T17:29:13.120 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:13.121 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:13.121 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:13.121 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:12 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:13.121 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:12 vm11 ceph-mon[53973]: pgmap v79: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:12 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:12 vm10 ceph-mon[53712]: pgmap v79: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:13.779 INFO:teuthology.orchestra.run.vm11.stdout:Created osd(s) 7 on host 'vm11' 2026-03-09T17:29:14.086 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:13 vm11 ceph-mon[53973]: from='osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T17:29:14.086 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:13 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:14.110 DEBUG:teuthology.orchestra.run.vm11:osd.7> sudo journalctl -f -n 0 -u ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea@osd.7.service 2026-03-09T17:29:14.115 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-09T17:29:14.115 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd stat -f json 2026-03-09T17:29:14.286 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:13 vm10 ceph-mon[53712]: from='osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-09T17:29:14.286 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:13 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:14.623 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:14.818 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":40,"num_osds":8,"num_up_osds":7,"osd_up_since":1773077345,"num_in_osds":8,"osd_in_since":1773077345,"num_remapped_pgs":0} 2026-03-09T17:29:15.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:15 vm11 ceph-mon[53973]: from='osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-09T17:29:15.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:15 vm11 ceph-mon[53973]: osdmap e40: 8 total, 7 up, 8 in 2026-03-09T17:29:15.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:15 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:15.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:15 vm11 ceph-mon[53973]: from='osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:29:15.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:15 vm11 ceph-mon[53973]: pgmap v81: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:15.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:15 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:15.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:15 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/907124273' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:29:15.341 INFO:journalctl@ceph.osd.7.vm11.stdout:Mar 09 17:29:15 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-7[66386]: 2026-03-09T17:29:15.033+0000 7f45dba56700 -1 osd.7 0 waiting for initial osdmap 2026-03-09T17:29:15.341 INFO:journalctl@ceph.osd.7.vm11.stdout:Mar 09 17:29:15 vm11 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-osd-7[66386]: 2026-03-09T17:29:15.050+0000 7f45d6bee700 -1 osd.7 41 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T17:29:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:15 vm10 ceph-mon[53712]: from='osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441]' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-09T17:29:15.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:15 vm10 ceph-mon[53712]: osdmap e40: 8 total, 7 up, 8 in 2026-03-09T17:29:15.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:15 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:15.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:15 vm10 ceph-mon[53712]: from='osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T17:29:15.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:15 vm10 ceph-mon[53712]: pgmap v81: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:15.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:15 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:15.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:15 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/907124273' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:29:15.819 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd stat -f json 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: osdmap e41: 8 total, 7 up, 8 in 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:16.113 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:16 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441]' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T17:29:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: osdmap e41: 8 total, 7 up, 8 in 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:16.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:16 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:16.348 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:16.422 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":42,"num_osds":8,"num_up_osds":8,"osd_up_since":1773077356,"num_in_osds":8,"osd_in_since":1773077345,"num_remapped_pgs":1} 2026-03-09T17:29:16.423 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd dump --format=json 2026-03-09T17:29:16.590 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:16.966 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:16.966 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":42,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","created":"2026-03-09T17:26:15.226580+0000","modified":"2026-03-09T17:29:16.025110+0000","last_up_change":"2026-03-09T17:29:16.025110+0000","last_in_change":"2026-03-09T17:29:05.941189+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T17:28:28.248007+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"6ac80623-9e73-4d06-ab1e-44875a280f2c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":41,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6803","nonce":1156255049}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6805","nonce":1156255049}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6809","nonce":1156255049}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6807","nonce":1156255049}]},"public_addr":"192.168.123.110:6803/1156255049","cluster_addr":"192.168.123.110:6805/1156255049","heartbeat_back_addr":"192.168.123.110:6809/1156255049","heartbeat_front_addr":"192.168.123.110:6807/1156255049","state":["exists","up"]},{"osd":1,"uuid":"fc09d5ce-67c5-4201-a2f2-fe1cb3b21fad","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":28,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6811","nonce":1744031210}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6813","nonce":1744031210}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6816","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6817","nonce":1744031210}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6815","nonce":1744031210}]},"public_addr":"192.168.123.110:6811/1744031210","cluster_addr":"192.168.123.110:6813/1744031210","heartbeat_back_addr":"192.168.123.110:6817/1744031210","heartbeat_front_addr":"192.168.123.110:6815/1744031210","state":["exists","up"]},{"osd":2,"uuid":"9de17247-790c-41ec-9cd7-a62e65ec8735","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6818","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6819","nonce":3248441350}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6820","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6821","nonce":3248441350}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6824","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6825","nonce":3248441350}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6822","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6823","nonce":3248441350}]},"public_addr":"192.168.123.110:6819/3248441350","cluster_addr":"192.168.123.110:6821/3248441350","heartbeat_back_addr":"192.168.123.110:6825/3248441350","heartbeat_front_addr":"192.168.123.110:6823/3248441350","state":["exists","up"]},{"osd":3,"uuid":"5b05874b-c02b-475c-8aa7-e90fa7ea69b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6826","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6827","nonce":2088446903}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6828","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6829","nonce":2088446903}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6832","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6833","nonce":2088446903}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6830","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6831","nonce":2088446903}]},"public_addr":"192.168.123.110:6827/2088446903","cluster_addr":"192.168.123.110:6829/2088446903","heartbeat_back_addr":"192.168.123.110:6833/2088446903","heartbeat_front_addr":"192.168.123.110:6831/2088446903","state":["exists","up"]},{"osd":4,"uuid":"ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6800","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6801","nonce":236530777}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6802","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6803","nonce":236530777}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6806","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6807","nonce":236530777}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6804","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6805","nonce":236530777}]},"public_addr":"192.168.123.111:6801/236530777","cluster_addr":"192.168.123.111:6803/236530777","heartbeat_back_addr":"192.168.123.111:6807/236530777","heartbeat_front_addr":"192.168.123.111:6805/236530777","state":["exists","up"]},{"osd":5,"uuid":"015dd696-8386-4fa0-b5f2-f22dc317f27f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":33,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6808","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6809","nonce":2564328090}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6810","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6811","nonce":2564328090}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6814","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6815","nonce":2564328090}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6812","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6813","nonce":2564328090}]},"public_addr":"192.168.123.111:6809/2564328090","cluster_addr":"192.168.123.111:6811/2564328090","heartbeat_back_addr":"192.168.123.111:6815/2564328090","heartbeat_front_addr":"192.168.123.111:6813/2564328090","state":["exists","up"]},{"osd":6,"uuid":"ac005d54-a0d8-4362-874e-656309d25f84","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":38,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6816","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6817","nonce":605495969}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6818","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6819","nonce":605495969}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6822","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6823","nonce":605495969}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6820","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6821","nonce":605495969}]},"public_addr":"192.168.123.111:6817/605495969","cluster_addr":"192.168.123.111:6819/605495969","heartbeat_back_addr":"192.168.123.111:6823/605495969","heartbeat_front_addr":"192.168.123.111:6821/605495969","state":["exists","up"]},{"osd":7,"uuid":"eedd693b-c665-4416-af3c-40637ce85fc7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":42,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6824","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6825","nonce":4145509441}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6826","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6827","nonce":4145509441}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6830","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6831","nonce":4145509441}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6828","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6829","nonce":4145509441}]},"public_addr":"192.168.123.111:6825/4145509441","cluster_addr":"192.168.123.111:6827/4145509441","heartbeat_back_addr":"192.168.123.111:6831/4145509441","heartbeat_front_addr":"192.168.123.111:6829/4145509441","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:06.866443+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:16.315690+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:25.727527+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:34.931851+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:44.742761+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:54.143720+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:29:04.720189+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[{"pgid":"1.0","osds":[0,6,1]}],"primary_temp":[],"blocklist":{"192.168.123.110:6800/1633567637":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/652975758":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/3422950029":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/3654887418":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/3969455737":"2026-03-10T17:26:43.832549+0000","192.168.123.110:6801/1633567637":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/90493270":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/3592193872":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6800/447343185":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6801/447343185":"2026-03-10T17:26:30.710330+0000","192.168.123.110:0/1488239144":"2026-03-10T17:27:27.807090+0000","192.168.123.110:6800/1362225250":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/111598728":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6801/1362225250":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/1785977233":"2026-03-10T17:26:30.710330+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T17:29:17.057 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-09T17:28:28.248007+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '20', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}}] 2026-03-09T17:29:17.057 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd pool get .mgr pg_num 2026-03-09T17:29:17.236 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: purged_snaps scrub starts 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: purged_snaps scrub ok 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: Detected new or changed devices on vm11 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: pgmap v84: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441] boot 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: osdmap e42: 8 total, 8 up, 8 in 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/1187919773' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:29:17.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:17 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/3620693010' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: purged_snaps scrub starts 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: purged_snaps scrub ok 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: Detected new or changed devices on vm11 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: pgmap v84: 1 pgs: 1 active+recovering; 449 KiB data, 41 MiB used, 140 GiB / 140 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: osd.7 [v2:192.168.123.111:6824/4145509441,v1:192.168.123.111:6825/4145509441] boot 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: osdmap e42: 8 total, 8 up, 8 in 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/1187919773' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T17:29:17.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:17 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/3620693010' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:29:17.958 INFO:teuthology.orchestra.run.vm10.stdout:pg_num: 1 2026-03-09T17:29:18.064 INFO:tasks.cephadm:Setting up client nodes... 2026-03-09T17:29:18.064 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-09T17:29:18.657 INFO:teuthology.orchestra.run.vm10.stdout:[client.0] 2026-03-09T17:29:18.657 INFO:teuthology.orchestra.run.vm10.stdout: key = AQBuA69pPfl+JhAAq/qAApCqplE+sbTHnhcN7w== 2026-03-09T17:29:18.726 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T17:29:18.726 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-09T17:29:18.726 DEBUG:teuthology.orchestra.run.vm10:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-09T17:29:18.771 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-09T17:29:18.771 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-09T17:29:18.772 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph mgr dump --format=json 2026-03-09T17:29:18.840 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:18 vm10 ceph-mon[53712]: osdmap e43: 8 total, 8 up, 8 in 2026-03-09T17:29:18.840 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:18 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/3012914602' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T17:29:18.840 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:18 vm10 ceph-mon[53712]: pgmap v86: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:18.840 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:18 vm10 ceph-mon[53712]: osdmap e44: 8 total, 8 up, 8 in 2026-03-09T17:29:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:18 vm11 ceph-mon[53973]: osdmap e43: 8 total, 8 up, 8 in 2026-03-09T17:29:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:18 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/3012914602' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T17:29:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:18 vm11 ceph-mon[53973]: pgmap v86: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:18 vm11 ceph-mon[53973]: osdmap e44: 8 total, 8 up, 8 in 2026-03-09T17:29:18.994 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:19.454 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:19.712 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:19 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/4013545710' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:29:19.712 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:19 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/4013545710' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:29:19.712 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:19 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/915989528' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T17:29:19.713 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":19,"active_gid":14232,"active_name":"vm10.qompzp","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":2416823152},{"type":"v1","addr":"192.168.123.110:6801","nonce":2416823152}]},"active_addr":"192.168.123.110:6801/2416823152","active_change":"2026-03-09T17:27:27.807727+0000","active_mgr_features":4540138303579357183,"available":true,"standbys":[{"gid":14252,"name":"vm11.ohhvcs","mgr_features":4540138303579357183,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2400","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"7","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","upmap"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.23.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/ceph-grafana:8.3.5","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"docker.io/library/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"docker.io/arcts/keepalived","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.3.1","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.33.4","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"docker.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"noautoscale":{"name":"noautoscale","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"global autoscale flag","long_desc":"Option to turn on/off the autoscaler for all pools","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"drive_group_interval":{"name":"drive_group_interval","type":"float","level":"advanced","flags":0,"default_value":"300.0","min":"","max":"","enum_allowed":[],"desc":"interval in seconds between re-application of applied drive_groups","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2400","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"7","min":"0","max":"7","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 or 7 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","upmap"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.23.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/ceph-grafana:8.3.5","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"docker.io/library/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"docker.io/arcts/keepalived","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.3.1","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.33.4","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"docker.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"noautoscale":{"name":"noautoscale","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"global autoscale flag","long_desc":"Option to turn on/off the autoscaler for all pools","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"drive_group_interval":{"name":"drive_group_interval","type":"float","level":"advanced","flags":0,"default_value":"300.0","min":"","max":"","enum_allowed":[],"desc":"interval in seconds between re-application of applied drive_groups","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.110:8443/","prometheus":"http://192.168.123.110:9283/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"last_failure_osd_epoch":5,"active_clients":[{"addrvec":[{"type":"v2","addr":"192.168.123.110:0","nonce":1118170747}]},{"addrvec":[{"type":"v2","addr":"192.168.123.110:0","nonce":704581483}]},{"addrvec":[{"type":"v2","addr":"192.168.123.110:0","nonce":1945113616}]},{"addrvec":[{"type":"v2","addr":"192.168.123.110:0","nonce":3456634729}]}]}} 2026-03-09T17:29:19.714 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-09T17:29:19.714 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-09T17:29:19.714 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd dump --format=json 2026-03-09T17:29:19.944 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:19 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/4013545710' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-09T17:29:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:19 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/4013545710' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-09T17:29:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:19 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/915989528' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T17:29:20.411 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:20.411 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":44,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","created":"2026-03-09T17:26:15.226580+0000","modified":"2026-03-09T17:29:18.184773+0000","last_up_change":"2026-03-09T17:29:16.025110+0000","last_in_change":"2026-03-09T17:29:05.941189+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T17:28:28.248007+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"6ac80623-9e73-4d06-ab1e-44875a280f2c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":42,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6803","nonce":1156255049}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6805","nonce":1156255049}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6809","nonce":1156255049}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6807","nonce":1156255049}]},"public_addr":"192.168.123.110:6803/1156255049","cluster_addr":"192.168.123.110:6805/1156255049","heartbeat_back_addr":"192.168.123.110:6809/1156255049","heartbeat_front_addr":"192.168.123.110:6807/1156255049","state":["exists","up"]},{"osd":1,"uuid":"fc09d5ce-67c5-4201-a2f2-fe1cb3b21fad","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":28,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6811","nonce":1744031210}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6813","nonce":1744031210}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6816","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6817","nonce":1744031210}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6815","nonce":1744031210}]},"public_addr":"192.168.123.110:6811/1744031210","cluster_addr":"192.168.123.110:6813/1744031210","heartbeat_back_addr":"192.168.123.110:6817/1744031210","heartbeat_front_addr":"192.168.123.110:6815/1744031210","state":["exists","up"]},{"osd":2,"uuid":"9de17247-790c-41ec-9cd7-a62e65ec8735","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6818","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6819","nonce":3248441350}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6820","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6821","nonce":3248441350}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6824","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6825","nonce":3248441350}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6822","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6823","nonce":3248441350}]},"public_addr":"192.168.123.110:6819/3248441350","cluster_addr":"192.168.123.110:6821/3248441350","heartbeat_back_addr":"192.168.123.110:6825/3248441350","heartbeat_front_addr":"192.168.123.110:6823/3248441350","state":["exists","up"]},{"osd":3,"uuid":"5b05874b-c02b-475c-8aa7-e90fa7ea69b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6826","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6827","nonce":2088446903}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6828","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6829","nonce":2088446903}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6832","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6833","nonce":2088446903}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6830","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6831","nonce":2088446903}]},"public_addr":"192.168.123.110:6827/2088446903","cluster_addr":"192.168.123.110:6829/2088446903","heartbeat_back_addr":"192.168.123.110:6833/2088446903","heartbeat_front_addr":"192.168.123.110:6831/2088446903","state":["exists","up"]},{"osd":4,"uuid":"ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6800","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6801","nonce":236530777}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6802","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6803","nonce":236530777}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6806","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6807","nonce":236530777}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6804","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6805","nonce":236530777}]},"public_addr":"192.168.123.111:6801/236530777","cluster_addr":"192.168.123.111:6803/236530777","heartbeat_back_addr":"192.168.123.111:6807/236530777","heartbeat_front_addr":"192.168.123.111:6805/236530777","state":["exists","up"]},{"osd":5,"uuid":"015dd696-8386-4fa0-b5f2-f22dc317f27f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":33,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6808","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6809","nonce":2564328090}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6810","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6811","nonce":2564328090}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6814","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6815","nonce":2564328090}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6812","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6813","nonce":2564328090}]},"public_addr":"192.168.123.111:6809/2564328090","cluster_addr":"192.168.123.111:6811/2564328090","heartbeat_back_addr":"192.168.123.111:6815/2564328090","heartbeat_front_addr":"192.168.123.111:6813/2564328090","state":["exists","up"]},{"osd":6,"uuid":"ac005d54-a0d8-4362-874e-656309d25f84","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":38,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6816","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6817","nonce":605495969}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6818","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6819","nonce":605495969}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6822","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6823","nonce":605495969}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6820","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6821","nonce":605495969}]},"public_addr":"192.168.123.111:6817/605495969","cluster_addr":"192.168.123.111:6819/605495969","heartbeat_back_addr":"192.168.123.111:6823/605495969","heartbeat_front_addr":"192.168.123.111:6821/605495969","state":["exists","up"]},{"osd":7,"uuid":"eedd693b-c665-4416-af3c-40637ce85fc7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":42,"up_thru":43,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6824","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6825","nonce":4145509441}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6826","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6827","nonce":4145509441}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6830","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6831","nonce":4145509441}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6828","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6829","nonce":4145509441}]},"public_addr":"192.168.123.111:6825/4145509441","cluster_addr":"192.168.123.111:6827/4145509441","heartbeat_back_addr":"192.168.123.111:6831/4145509441","heartbeat_front_addr":"192.168.123.111:6829/4145509441","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:06.866443+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:16.315690+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:25.727527+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:34.931851+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:44.742761+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:54.143720+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:29:04.720189+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:29:14.420572+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.110:6800/1633567637":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/652975758":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/3422950029":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/3654887418":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/3969455737":"2026-03-10T17:26:43.832549+0000","192.168.123.110:6801/1633567637":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/90493270":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/3592193872":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6800/447343185":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6801/447343185":"2026-03-10T17:26:30.710330+0000","192.168.123.110:0/1488239144":"2026-03-10T17:27:27.807090+0000","192.168.123.110:6800/1362225250":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/111598728":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6801/1362225250":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/1785977233":"2026-03-10T17:26:30.710330+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T17:29:20.505 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-09T17:29:20.505 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd dump --format=json 2026-03-09T17:29:20.672 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:20.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:20 vm10 ceph-mon[53712]: pgmap v88: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:20.751 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:20 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/3009912149' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:29:21.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:20 vm11 ceph-mon[53973]: pgmap v88: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:21.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:20 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/3009912149' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:29:21.097 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:21.097 INFO:teuthology.orchestra.run.vm10.stdout:{"epoch":44,"fsid":"01c7db6a-1bdd-11f1-ac39-b134231210ea","created":"2026-03-09T17:26:15.226580+0000","modified":"2026-03-09T17:29:18.184773+0000","last_up_change":"2026-03-09T17:29:16.025110+0000","last_in_change":"2026-03-09T17:29:05.941189+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":18,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"quincy","pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T17:28:28.248007+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}}}],"osds":[{"osd":0,"uuid":"6ac80623-9e73-4d06-ab1e-44875a280f2c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":9,"up_thru":42,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6803","nonce":1156255049}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6805","nonce":1156255049}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6809","nonce":1156255049}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1156255049},{"type":"v1","addr":"192.168.123.110:6807","nonce":1156255049}]},"public_addr":"192.168.123.110:6803/1156255049","cluster_addr":"192.168.123.110:6805/1156255049","heartbeat_back_addr":"192.168.123.110:6809/1156255049","heartbeat_front_addr":"192.168.123.110:6807/1156255049","state":["exists","up"]},{"osd":1,"uuid":"fc09d5ce-67c5-4201-a2f2-fe1cb3b21fad","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":28,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6811","nonce":1744031210}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6813","nonce":1744031210}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6816","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6817","nonce":1744031210}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":1744031210},{"type":"v1","addr":"192.168.123.110:6815","nonce":1744031210}]},"public_addr":"192.168.123.110:6811/1744031210","cluster_addr":"192.168.123.110:6813/1744031210","heartbeat_back_addr":"192.168.123.110:6817/1744031210","heartbeat_front_addr":"192.168.123.110:6815/1744031210","state":["exists","up"]},{"osd":2,"uuid":"9de17247-790c-41ec-9cd7-a62e65ec8735","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6818","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6819","nonce":3248441350}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6820","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6821","nonce":3248441350}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6824","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6825","nonce":3248441350}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6822","nonce":3248441350},{"type":"v1","addr":"192.168.123.110:6823","nonce":3248441350}]},"public_addr":"192.168.123.110:6819/3248441350","cluster_addr":"192.168.123.110:6821/3248441350","heartbeat_back_addr":"192.168.123.110:6825/3248441350","heartbeat_front_addr":"192.168.123.110:6823/3248441350","state":["exists","up"]},{"osd":3,"uuid":"5b05874b-c02b-475c-8aa7-e90fa7ea69b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6826","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6827","nonce":2088446903}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6828","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6829","nonce":2088446903}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6832","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6833","nonce":2088446903}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6830","nonce":2088446903},{"type":"v1","addr":"192.168.123.110:6831","nonce":2088446903}]},"public_addr":"192.168.123.110:6827/2088446903","cluster_addr":"192.168.123.110:6829/2088446903","heartbeat_back_addr":"192.168.123.110:6833/2088446903","heartbeat_front_addr":"192.168.123.110:6831/2088446903","state":["exists","up"]},{"osd":4,"uuid":"ffe8a8e7-c2ef-4b41-b3de-449108f9c5fa","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6800","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6801","nonce":236530777}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6802","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6803","nonce":236530777}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6806","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6807","nonce":236530777}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6804","nonce":236530777},{"type":"v1","addr":"192.168.123.111:6805","nonce":236530777}]},"public_addr":"192.168.123.111:6801/236530777","cluster_addr":"192.168.123.111:6803/236530777","heartbeat_back_addr":"192.168.123.111:6807/236530777","heartbeat_front_addr":"192.168.123.111:6805/236530777","state":["exists","up"]},{"osd":5,"uuid":"015dd696-8386-4fa0-b5f2-f22dc317f27f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":33,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6808","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6809","nonce":2564328090}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6810","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6811","nonce":2564328090}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6814","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6815","nonce":2564328090}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6812","nonce":2564328090},{"type":"v1","addr":"192.168.123.111:6813","nonce":2564328090}]},"public_addr":"192.168.123.111:6809/2564328090","cluster_addr":"192.168.123.111:6811/2564328090","heartbeat_back_addr":"192.168.123.111:6815/2564328090","heartbeat_front_addr":"192.168.123.111:6813/2564328090","state":["exists","up"]},{"osd":6,"uuid":"ac005d54-a0d8-4362-874e-656309d25f84","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":38,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6816","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6817","nonce":605495969}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6818","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6819","nonce":605495969}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6822","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6823","nonce":605495969}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6820","nonce":605495969},{"type":"v1","addr":"192.168.123.111:6821","nonce":605495969}]},"public_addr":"192.168.123.111:6817/605495969","cluster_addr":"192.168.123.111:6819/605495969","heartbeat_back_addr":"192.168.123.111:6823/605495969","heartbeat_front_addr":"192.168.123.111:6821/605495969","state":["exists","up"]},{"osd":7,"uuid":"eedd693b-c665-4416-af3c-40637ce85fc7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":42,"up_thru":43,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6824","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6825","nonce":4145509441}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6826","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6827","nonce":4145509441}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6830","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6831","nonce":4145509441}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6828","nonce":4145509441},{"type":"v1","addr":"192.168.123.111:6829","nonce":4145509441}]},"public_addr":"192.168.123.111:6825/4145509441","cluster_addr":"192.168.123.111:6827/4145509441","heartbeat_back_addr":"192.168.123.111:6831/4145509441","heartbeat_front_addr":"192.168.123.111:6829/4145509441","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:06.866443+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:16.315690+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:25.727527+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:34.931851+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:44.742761+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:28:54.143720+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:29:04.720189+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540138303579357183,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T17:29:14.420572+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.110:6800/1633567637":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/652975758":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/3422950029":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/3654887418":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/3969455737":"2026-03-10T17:26:43.832549+0000","192.168.123.110:6801/1633567637":"2026-03-10T17:27:27.807090+0000","192.168.123.110:0/90493270":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/3592193872":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6800/447343185":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6801/447343185":"2026-03-10T17:26:30.710330+0000","192.168.123.110:0/1488239144":"2026-03-10T17:27:27.807090+0000","192.168.123.110:6800/1362225250":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/111598728":"2026-03-10T17:26:30.710330+0000","192.168.123.110:6801/1362225250":"2026-03-10T17:26:43.832549+0000","192.168.123.110:0/1785977233":"2026-03-10T17:26:30.710330+0000"},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T17:29:21.242 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph tell osd.0 flush_pg_stats 2026-03-09T17:29:21.243 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph tell osd.1 flush_pg_stats 2026-03-09T17:29:21.243 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph tell osd.2 flush_pg_stats 2026-03-09T17:29:21.243 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph tell osd.3 flush_pg_stats 2026-03-09T17:29:21.243 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph tell osd.4 flush_pg_stats 2026-03-09T17:29:21.243 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph tell osd.5 flush_pg_stats 2026-03-09T17:29:21.243 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph tell osd.6 flush_pg_stats 2026-03-09T17:29:21.244 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph tell osd.7 flush_pg_stats 2026-03-09T17:29:21.631 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:21 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/2959427178' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:29:21.835 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:21.838 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:21.946 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:22.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:21 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/2959427178' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T17:29:22.193 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:22.194 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:22.326 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:22.343 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:22.346 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:23.271 INFO:teuthology.orchestra.run.vm10.stdout:73014444045 2026-03-09T17:29:23.271 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd last-stat-seq osd.2 2026-03-09T17:29:23.282 INFO:teuthology.orchestra.run.vm10.stdout:55834574863 2026-03-09T17:29:23.282 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd last-stat-seq osd.1 2026-03-09T17:29:23.287 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:23 vm10 ceph-mon[53712]: pgmap v89: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:23 vm11 ceph-mon[53973]: pgmap v89: 1 pgs: 1 active+recovering; 449 KiB data, 47 MiB used, 160 GiB / 160 GiB avail; 1/6 objects misplaced (16.667%) 2026-03-09T17:29:24.239 INFO:teuthology.orchestra.run.vm10.stdout:180388626435 2026-03-09T17:29:24.239 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd last-stat-seq osd.7 2026-03-09T17:29:24.265 INFO:teuthology.orchestra.run.vm10.stdout:98784247819 2026-03-09T17:29:24.265 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd last-stat-seq osd.3 2026-03-09T17:29:24.267 INFO:teuthology.orchestra.run.vm10.stdout:158913789957 2026-03-09T17:29:24.281 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd last-stat-seq osd.6 2026-03-09T17:29:24.366 INFO:teuthology.orchestra.run.vm10.stdout:38654705680 2026-03-09T17:29:24.367 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd last-stat-seq osd.0 2026-03-09T17:29:24.383 INFO:teuthology.orchestra.run.vm10.stdout:137438953479 2026-03-09T17:29:24.384 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd last-stat-seq osd.5 2026-03-09T17:29:24.435 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:24.514 INFO:teuthology.orchestra.run.vm10.stdout:115964117001 2026-03-09T17:29:24.515 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph osd last-stat-seq osd.4 2026-03-09T17:29:24.525 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:25.007 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:25.073 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:25.086 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:25 vm10 ceph-mon[53712]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s, 0 objects/s recovering 2026-03-09T17:29:25.331 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:25.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:25 vm11 ceph-mon[53973]: pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s, 0 objects/s recovering 2026-03-09T17:29:25.614 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:25.616 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:25.632 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:25.814 INFO:teuthology.orchestra.run.vm10.stdout:73014444045 2026-03-09T17:29:25.913 INFO:teuthology.orchestra.run.vm10.stdout:55834574864 2026-03-09T17:29:26.141 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:26 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/3870363175' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T17:29:26.141 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:26 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/873925159' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T17:29:26.181 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574863 got 55834574864 for osd.1 2026-03-09T17:29:26.181 DEBUG:teuthology.parallel:result is None 2026-03-09T17:29:26.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:26 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/3870363175' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T17:29:26.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:26 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/873925159' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T17:29:26.441 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444045 got 73014444045 for osd.2 2026-03-09T17:29:26.441 DEBUG:teuthology.parallel:result is None 2026-03-09T17:29:26.459 INFO:teuthology.orchestra.run.vm10.stdout:98784247819 2026-03-09T17:29:26.604 INFO:teuthology.orchestra.run.vm10.stdout:180388626435 2026-03-09T17:29:26.608 INFO:teuthology.orchestra.run.vm10.stdout:158913789957 2026-03-09T17:29:26.697 INFO:tasks.cephadm.ceph_manager.ceph:need seq 98784247819 got 98784247819 for osd.3 2026-03-09T17:29:26.697 DEBUG:teuthology.parallel:result is None 2026-03-09T17:29:26.986 INFO:tasks.cephadm.ceph_manager.ceph:need seq 158913789957 got 158913789957 for osd.6 2026-03-09T17:29:26.986 DEBUG:teuthology.parallel:result is None 2026-03-09T17:29:27.159 INFO:tasks.cephadm.ceph_manager.ceph:need seq 180388626435 got 180388626435 for osd.7 2026-03-09T17:29:27.159 DEBUG:teuthology.parallel:result is None 2026-03-09T17:29:27.224 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:27 vm10 ceph-mon[53712]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s, 0 objects/s recovering 2026-03-09T17:29:27.224 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:27 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/4192241988' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T17:29:27.224 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:27 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/1704437483' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-09T17:29:27.224 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:27 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/1001823043' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T17:29:27.461 INFO:teuthology.orchestra.run.vm10.stdout:115964117001 2026-03-09T17:29:27.478 INFO:teuthology.orchestra.run.vm10.stdout:137438953479 2026-03-09T17:29:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:27 vm11 ceph-mon[53973]: pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s, 0 objects/s recovering 2026-03-09T17:29:27.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:27 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/4192241988' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T17:29:27.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:27 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/1704437483' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-09T17:29:27.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:27 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/1001823043' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-09T17:29:27.594 INFO:teuthology.orchestra.run.vm10.stdout:38654705681 2026-03-09T17:29:27.724 INFO:tasks.cephadm.ceph_manager.ceph:need seq 115964117001 got 115964117001 for osd.4 2026-03-09T17:29:27.724 DEBUG:teuthology.parallel:result is None 2026-03-09T17:29:28.021 INFO:tasks.cephadm.ceph_manager.ceph:need seq 38654705680 got 38654705681 for osd.0 2026-03-09T17:29:28.021 DEBUG:teuthology.parallel:result is None 2026-03-09T17:29:28.155 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:28 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/4290132452' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T17:29:28.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:28 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/809688824' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T17:29:28.156 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:28 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/3975811742' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T17:29:28.156 INFO:tasks.cephadm.ceph_manager.ceph:need seq 137438953479 got 137438953479 for osd.5 2026-03-09T17:29:28.156 DEBUG:teuthology.parallel:result is None 2026-03-09T17:29:28.157 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-09T17:29:28.157 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph pg dump --format=json 2026-03-09T17:29:28.312 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:28.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:28 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/4290132452' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T17:29:28.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:28 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/809688824' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T17:29:28.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:28 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/3975811742' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T17:29:28.774 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:28.780 INFO:teuthology.orchestra.run.vm10.stderr:dumped all 2026-03-09T17:29:28.872 INFO:teuthology.orchestra.run.vm10.stdout:{"pg_ready":true,"pg_map":{"version":92,"stamp":"2026-03-09T17:29:28.034870+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":5,"num_osds":8,"num_per_pool_osds":4,"num_per_pool_omap_osds":4,"kb":167739392,"kb_used":48712,"kb_used_data":4616,"kb_used_omap":0,"kb_used_meta":44032,"kb_avail":167690680,"statfs":{"total":171765137408,"available":171715256320,"internally_reserved":0,"allocated":4726784,"data_stored":2623524,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45088768},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":-1,"num_objects_missing":-1,"num_objects_degraded":0,"num_objects_misplaced":-1,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":327680,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"10.001642"},"pg_stats":[{"pgid":"1.0","version":"20'76","reported_seq":16,"reported_epoch":44,"state":"active+clean","last_fresh":"2026-03-09T17:29:18.514436+0000","last_change":"2026-03-09T17:29:18.514436+0000","last_active":"2026-03-09T17:29:18.514436+0000","last_peered":"2026-03-09T17:29:18.514436+0000","last_clean":"2026-03-09T17:29:18.514436+0000","last_became_active":"2026-03-09T17:29:18.193784+0000","last_became_peered":"2026-03-09T17:29:18.193784+0000","last_unstale":"2026-03-09T17:29:18.514436+0000","last_undegraded":"2026-03-09T17:29:18.514436+0000","last_fullsized":"2026-03-09T17:29:18.514436+0000","mapping_epoch":43,"log_start":"0'0","ondisk_log_start":"0'0","created":18,"last_epoch_clean":44,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:28:28.296956+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:28:28.296956+0000","last_clean_scrub_stamp":"2026-03-09T17:28:28.296956+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T04:41:14.899810+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1204224,"data_stored":1193520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":42,"seq":180388626436,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6076,"kb_used_data":820,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961348,"statfs":{"total":21470642176,"available":21464420352,"internally_reserved":0,"allocated":839680,"data_stored":576433,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60199999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63300000000000001}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.74199999999999999}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.93000000000000005}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.91400000000000003}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57899999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.0049999999999999}]}]},{"osd":6,"up_from":37,"seq":158913789958,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6080,"kb_used_data":824,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961344,"statfs":{"total":21470642176,"available":21464416256,"internally_reserved":0,"allocated":843776,"data_stored":576748,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.73099999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70099999999999996}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.78100000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65200000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.746}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.78800000000000003}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.98399999999999999}]}]},{"osd":1,"up_from":13,"seq":55834574864,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6324,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20961100,"statfs":{"total":21470642176,"available":21464166400,"internally_reserved":0,"allocated":438272,"data_stored":178593,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Mon Mar 9 17:29:17 2026","interfaces":[{"interface":"back","average":{"1min":0.498,"5min":0.498,"15min":0.498},"min":{"1min":0.23100000000000001,"5min":0.23100000000000001,"15min":0.23100000000000001},"max":{"1min":0.91200000000000003,"5min":0.91200000000000003,"15min":0.91200000000000003},"last":0.66400000000000003},{"interface":"front","average":{"1min":0.80600000000000005,"5min":0.80600000000000005,"15min":0.80600000000000005},"min":{"1min":0.26900000000000002,"5min":0.26900000000000002,"15min":0.26900000000000002},"max":{"1min":4.4240000000000004,"5min":4.4240000000000004,"15min":4.4240000000000004},"last":0.73699999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60799999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.77300000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79600000000000004}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82199999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.98599999999999999}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86499999999999999}]}]},{"osd":0,"up_from":9,"seq":38654705681,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6720,"kb_used_data":824,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960704,"statfs":{"total":21470642176,"available":21463760896,"internally_reserved":0,"allocated":843776,"data_stored":576748,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Mon Mar 9 17:29:18 2026","interfaces":[{"interface":"back","average":{"1min":0.51700000000000002,"5min":0.51700000000000002,"15min":0.51700000000000002},"min":{"1min":0.216,"5min":0.216,"15min":0.216},"max":{"1min":0.92100000000000004,"5min":0.92100000000000004,"15min":0.92100000000000004},"last":0.60599999999999998},{"interface":"front","average":{"1min":0.57099999999999995,"5min":0.57099999999999995,"15min":0.57099999999999995},"min":{"1min":0.252,"5min":0.252,"15min":0.252},"max":{"1min":0.93500000000000005,"5min":0.93500000000000005,"15min":0.93500000000000005},"last":0.54300000000000004}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58599999999999997}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53600000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52000000000000002}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57699999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59099999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59899999999999998}]}]},{"osd":2,"up_from":17,"seq":73014444046,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6264,"kb_used_data":432,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961160,"statfs":{"total":21470642176,"available":21464227840,"internally_reserved":0,"allocated":442368,"data_stored":178908,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60299999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72999999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70699999999999996}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.91900000000000004}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67300000000000004}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.98199999999999998}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86899999999999999}]}]},{"osd":3,"up_from":23,"seq":98784247820,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5752,"kb_used_data":432,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961672,"statfs":{"total":21470642176,"available":21464752128,"internally_reserved":0,"allocated":442368,"data_stored":178908,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.054}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.016}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70799999999999996}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.438}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.98399999999999999}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.40100000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.998}]}]},{"osd":4,"up_from":27,"seq":115964117002,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":178593,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":4.101}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86899999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":3.7679999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":3.9670000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86399999999999999}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.48699999999999999}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.84699999999999998}]}]},{"osd":5,"up_from":32,"seq":137438953480,"num_pgs":1,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":178593,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.89500000000000002}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.88200000000000001}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.93200000000000005}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.91700000000000004}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67200000000000004}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65200000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60799999999999998}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T17:29:28.873 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph pg dump --format=json 2026-03-09T17:29:29.053 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:29.388 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:29 vm10 ceph-mon[53712]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s, 0 objects/s recovering 2026-03-09T17:29:29.388 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:29:29.388 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:29:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:29 vm11 ceph-mon[53973]: pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s, 0 objects/s recovering 2026-03-09T17:29:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:29 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:29:29.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:29 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:29:29.627 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:29.630 INFO:teuthology.orchestra.run.vm10.stderr:dumped all 2026-03-09T17:29:29.833 INFO:teuthology.orchestra.run.vm10.stdout:{"pg_ready":true,"pg_map":{"version":92,"stamp":"2026-03-09T17:29:28.034870+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":5,"num_osds":8,"num_per_pool_osds":4,"num_per_pool_omap_osds":4,"kb":167739392,"kb_used":48712,"kb_used_data":4616,"kb_used_omap":0,"kb_used_meta":44032,"kb_avail":167690680,"statfs":{"total":171765137408,"available":171715256320,"internally_reserved":0,"allocated":4726784,"data_stored":2623524,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":45088768},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":-1,"num_objects_missing":-1,"num_objects_degraded":0,"num_objects_misplaced":-1,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":1,"num_bytes_recovered":327680,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"10.001642"},"pg_stats":[{"pgid":"1.0","version":"20'76","reported_seq":16,"reported_epoch":44,"state":"active+clean","last_fresh":"2026-03-09T17:29:18.514436+0000","last_change":"2026-03-09T17:29:18.514436+0000","last_active":"2026-03-09T17:29:18.514436+0000","last_peered":"2026-03-09T17:29:18.514436+0000","last_clean":"2026-03-09T17:29:18.514436+0000","last_became_active":"2026-03-09T17:29:18.193784+0000","last_became_peered":"2026-03-09T17:29:18.193784+0000","last_unstale":"2026-03-09T17:29:18.514436+0000","last_undegraded":"2026-03-09T17:29:18.514436+0000","last_fullsized":"2026-03-09T17:29:18.514436+0000","mapping_epoch":43,"log_start":"0'0","ondisk_log_start":"0'0","created":18,"last_epoch_clean":44,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T17:28:28.296956+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T17:28:28.296956+0000","last_clean_scrub_stamp":"2026-03-09T17:28:28.296956+0000","objects_scrubbed":0,"log_size":76,"ondisk_log_size":76,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T04:41:14.899810+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0,6],"acting":[7,0,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":96,"num_read_kb":82,"num_write":113,"num_write_kb":1372,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":397840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1204224,"data_stored":1193520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":76,"ondisk_log_size":76,"up":3,"acting":3,"num_store_stats":7}],"osd_stats":[{"osd":7,"up_from":42,"seq":180388626436,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6076,"kb_used_data":820,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961348,"statfs":{"total":21470642176,"available":21464420352,"internally_reserved":0,"allocated":839680,"data_stored":576433,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60199999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.63300000000000001}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.74199999999999999}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.93000000000000005}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.91400000000000003}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57899999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.0049999999999999}]}]},{"osd":6,"up_from":37,"seq":158913789958,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6080,"kb_used_data":824,"kb_used_omap":0,"kb_used_meta":5248,"kb_avail":20961344,"statfs":{"total":21470642176,"available":21464416256,"internally_reserved":0,"allocated":843776,"data_stored":576748,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5373952},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.73099999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70099999999999996}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.78100000000000003}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65200000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.746}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.78800000000000003}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.98399999999999999}]}]},{"osd":1,"up_from":13,"seq":55834574864,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6324,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20961100,"statfs":{"total":21470642176,"available":21464166400,"internally_reserved":0,"allocated":438272,"data_stored":178593,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Mon Mar 9 17:29:17 2026","interfaces":[{"interface":"back","average":{"1min":0.498,"5min":0.498,"15min":0.498},"min":{"1min":0.23100000000000001,"5min":0.23100000000000001,"15min":0.23100000000000001},"max":{"1min":0.91200000000000003,"5min":0.91200000000000003,"15min":0.91200000000000003},"last":0.66400000000000003},{"interface":"front","average":{"1min":0.80600000000000005,"5min":0.80600000000000005,"15min":0.80600000000000005},"min":{"1min":0.26900000000000002,"5min":0.26900000000000002,"15min":0.26900000000000002},"max":{"1min":4.4240000000000004,"5min":4.4240000000000004,"15min":4.4240000000000004},"last":0.73699999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60799999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.77300000000000002}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.79600000000000004}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.82199999999999995}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.98599999999999999}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86499999999999999}]}]},{"osd":0,"up_from":9,"seq":38654705681,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":6720,"kb_used_data":824,"kb_used_omap":0,"kb_used_meta":5888,"kb_avail":20960704,"statfs":{"total":21470642176,"available":21463760896,"internally_reserved":0,"allocated":843776,"data_stored":576748,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":6029312},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":1,"last update":"Mon Mar 9 17:29:18 2026","interfaces":[{"interface":"back","average":{"1min":0.51700000000000002,"5min":0.51700000000000002,"15min":0.51700000000000002},"min":{"1min":0.216,"5min":0.216,"15min":0.216},"max":{"1min":0.92100000000000004,"5min":0.92100000000000004,"15min":0.92100000000000004},"last":0.60599999999999998},{"interface":"front","average":{"1min":0.57099999999999995,"5min":0.57099999999999995,"15min":0.57099999999999995},"min":{"1min":0.252,"5min":0.252,"15min":0.252},"max":{"1min":0.93500000000000005,"5min":0.93500000000000005,"15min":0.93500000000000005},"last":0.54300000000000004}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.58599999999999997}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.53600000000000003}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.52000000000000002}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.57699999999999996}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59099999999999997}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.59899999999999998}]}]},{"osd":2,"up_from":17,"seq":73014444046,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":6264,"kb_used_data":432,"kb_used_omap":0,"kb_used_meta":5824,"kb_avail":20961160,"statfs":{"total":21470642176,"available":21464227840,"internally_reserved":0,"allocated":442368,"data_stored":178908,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5963776},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60299999999999998}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.72999999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70699999999999996}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.91900000000000004}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67300000000000004}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.98199999999999998}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86899999999999999}]}]},{"osd":3,"up_from":23,"seq":98784247820,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5752,"kb_used_data":432,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961672,"statfs":{"total":21470642176,"available":21464752128,"internally_reserved":0,"allocated":442368,"data_stored":178908,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.054}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":1.016}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.70799999999999996}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.438}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.98399999999999999}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.40100000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.998}]}]},{"osd":4,"up_from":27,"seq":115964117002,"num_pgs":0,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":178593,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":4.101}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86899999999999999}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":3.7679999999999998}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":3.9670000000000001}]},{"osd":5,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.86399999999999999}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.48699999999999999}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.84699999999999998}]}]},{"osd":5,"up_from":32,"seq":137438953480,"num_pgs":1,"num_osds":1,"num_per_pool_osds":0,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":5748,"kb_used_data":428,"kb_used_omap":0,"kb_used_meta":5312,"kb_avail":20961676,"statfs":{"total":21470642176,"available":21464756224,"internally_reserved":0,"allocated":438272,"data_stored":178593,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":5439488},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[{"osd":0,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.89500000000000002}]},{"osd":1,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.88200000000000001}]},{"osd":2,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.93200000000000005}]},{"osd":3,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.91700000000000004}]},{"osd":4,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.67200000000000004}]},{"osd":6,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.65200000000000002}]},{"osd":7,"last update":"Thu Jan 1 00:00:00 1970","interfaces":[{"interface":"back","average":{"1min":0,"5min":0,"15min":0},"min":{"1min":0,"5min":0,"15min":0},"max":{"1min":0,"5min":0,"15min":0},"last":0.60799999999999998}]}]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":401408,"data_stored":397840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T17:29:29.837 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-09T17:29:29.838 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-09T17:29:29.838 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-09T17:29:29.838 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph health --format=json 2026-03-09T17:29:30.010 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/01c7db6a-1bdd-11f1-ac39-b134231210ea/mon.vm10/config 2026-03-09T17:29:30.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:30 vm10 ceph-mon[53712]: from='client.14474 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:30.436 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:30.436 INFO:teuthology.orchestra.run.vm10.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-09T17:29:30.507 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-09T17:29:30.507 INFO:tasks.cephadm:Setup complete, yielding 2026-03-09T17:29:30.507 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T17:29:30.510 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm10.local 2026-03-09T17:29:30.510 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'ceph orch status' 2026-03-09T17:29:30.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:30 vm11 ceph-mon[53973]: from='client.14474 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:31.079 INFO:teuthology.orchestra.run.vm10.stdout:Backend: cephadm 2026-03-09T17:29:31.079 INFO:teuthology.orchestra.run.vm10.stdout:Available: Yes 2026-03-09T17:29:31.079 INFO:teuthology.orchestra.run.vm10.stdout:Paused: No 2026-03-09T17:29:31.177 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'ceph orch ps' 2026-03-09T17:29:31.354 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:31 vm10 ceph-mon[53712]: from='client.14478 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:31.354 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:31 vm10 ceph-mon[53712]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T17:29:31.354 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:31 vm10 ceph-mon[53712]: from='client.? 192.168.123.110:0/645129197' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T17:29:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:31 vm11 ceph-mon[53973]: from='client.14478 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:31.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:31 vm11 ceph-mon[53973]: pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T17:29:31.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:31 vm11 ceph-mon[53973]: from='client.? 192.168.123.110:0/645129197' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T17:29:31.679 INFO:teuthology.orchestra.run.vm10.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:alertmanager.vm10 vm10 *:9093,9094 running (89s) 57s ago 2m 18.1M - ba2b418f427c a3b6ec11d5da 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:crash.vm10 vm10 running (2m) 57s ago 2m 6949k - 17.2.0 e1d6a67b021e ceb0b3eba222 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:crash.vm11 vm11 running (119s) 17s ago 119s 7096k - 17.2.0 e1d6a67b021e ab464e9bc4ba 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:grafana.vm10 vm10 *:3000 running (86s) 57s ago 2m 42.9M - 8.3.5 dad864ee21e9 ce3e06cd108c 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:mgr.vm10.qompzp vm10 *:9283 running (3m) 57s ago 3m 457M - 17.2.0 e1d6a67b021e 55c714dc6c0b 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:mgr.vm11.ohhvcs vm11 *:8443,9283 running (118s) 17s ago 118s 418M - 17.2.0 e1d6a67b021e 86cac93b5a1f 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:mon.vm10 vm10 running (3m) 57s ago 3m 42.5M 2048M 17.2.0 e1d6a67b021e 8437e945a2e5 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:mon.vm11 vm11 running (117s) 17s ago 117s 43.2M 2048M 17.2.0 e1d6a67b021e 5d27ff0af313 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:node-exporter.vm10 vm10 *:9100 running (2m) 57s ago 2m 9877k - 1dbe0e931976 3b667ce3bdb2 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:node-exporter.vm11 vm11 *:9100 running (114s) 17s ago 114s 17.8M - 1dbe0e931976 170888f29265 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:osd.0 vm10 running (87s) 57s ago 87s 32.3M 4096M 17.2.0 e1d6a67b021e 796ebb5d7dde 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:osd.1 vm10 running (78s) 57s ago 77s 40.7M 4096M 17.2.0 e1d6a67b021e 10c2e1f48d98 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:osd.2 vm10 running (69s) 57s ago 69s 33.9M 4096M 17.2.0 e1d6a67b021e fc864a66361d 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:osd.3 vm10 running (59s) 57s ago 59s 11.9M 4096M 17.2.0 e1d6a67b021e 70c5b1bf49b4 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:osd.4 vm11 running (50s) 17s ago 50s 35.7M 4096M 17.2.0 e1d6a67b021e c892b4f9bdfe 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:osd.5 vm11 running (39s) 17s ago 39s 32.4M 4096M 17.2.0 e1d6a67b021e 26f33499def3 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:osd.6 vm11 running (30s) 17s ago 30s 31.7M 4096M 17.2.0 e1d6a67b021e c706b305f9cc 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:osd.7 vm11 running (19s) 17s ago 19s 15.7M 4096M 17.2.0 e1d6a67b021e 1da58fd1717e 2026-03-09T17:29:31.715 INFO:teuthology.orchestra.run.vm10.stdout:prometheus.vm10 vm10 *:9095 running (95s) 57s ago 95s 36.1M - 514e6a882f6e 521db3cdf8b1 2026-03-09T17:29:31.924 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'ceph orch ls' 2026-03-09T17:29:32.206 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:32 vm10 ceph-mon[53712]: from='client.14486 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:alertmanager ?:9093,9094 1/1 58s ago 2m count:1 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:crash 2/2 58s ago 2m * 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:grafana ?:3000 1/1 58s ago 2m count:1 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:mgr 2/2 58s ago 2m count:2 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:mon 2/2 58s ago 2m vm10:192.168.123.110=vm10;vm11:192.168.123.111=vm11;count:2 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:node-exporter ?:9100 2/2 58s ago 2m * 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:osd 8 58s ago - 2026-03-09T17:29:32.420 INFO:teuthology.orchestra.run.vm10.stdout:prometheus ?:9095 1/1 58s ago 2m count:1 2026-03-09T17:29:32.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:32 vm11 ceph-mon[53973]: from='client.14486 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:32.609 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'ceph orch host ls' 2026-03-09T17:29:33.443 INFO:teuthology.orchestra.run.vm10.stdout:HOST ADDR LABELS STATUS 2026-03-09T17:29:33.443 INFO:teuthology.orchestra.run.vm10.stdout:vm10 192.168.123.110 2026-03-09T17:29:33.443 INFO:teuthology.orchestra.run.vm10.stdout:vm11 192.168.123.111 2026-03-09T17:29:33.443 INFO:teuthology.orchestra.run.vm10.stdout:2 hosts in cluster 2026-03-09T17:29:33.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:33 vm10 ceph-mon[53712]: from='client.14490 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:33.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:33 vm10 ceph-mon[53712]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T17:29:33.784 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'ceph orch device ls' 2026-03-09T17:29:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:33 vm11 ceph-mon[53973]: from='client.14490 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:33.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:33 vm11 ceph-mon[53973]: pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REJECT REASONS 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:vm10 /dev/vdb hdd DWNBRSTVMM10001 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:vm10 /dev/vdc hdd DWNBRSTVMM10002 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:vm10 /dev/vdd hdd DWNBRSTVMM10003 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:vm10 /dev/vde hdd DWNBRSTVMM10004 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:vm11 /dev/vdb hdd DWNBRSTVMM11001 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:vm11 /dev/vdc hdd DWNBRSTVMM11002 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:vm11 /dev/vdd hdd DWNBRSTVMM11003 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T17:29:34.616 INFO:teuthology.orchestra.run.vm10.stdout:vm11 /dev/vde hdd DWNBRSTVMM11004 21.4G Insufficient space (<10 extents) on vgs, LVM detected, locked 2026-03-09T17:29:34.617 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:34 vm10 ceph-mon[53712]: from='client.14494 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:34.617 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:34 vm10 ceph-mon[53712]: from='client.14498 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:34.617 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:34 vm10 ceph-mon[53712]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T17:29:34.685 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-09T17:29:34.693 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm10.local 2026-03-09T17:29:34.693 DEBUG:teuthology.orchestra.run.vm10:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-09T17:29:34.729 INFO:teuthology.orchestra.run.vm10.stderr:+ systemctl stop nfs-server 2026-03-09T17:29:34.737 INFO:tasks.vip:Running commands on role host.b host ubuntu@vm11.local 2026-03-09T17:29:34.737 DEBUG:teuthology.orchestra.run.vm11:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-09T17:29:34.762 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:34 vm11 ceph-mon[53973]: from='client.14494 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:34.762 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:34 vm11 ceph-mon[53973]: from='client.14498 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:34.762 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:34 vm11 ceph-mon[53973]: pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s, 0 objects/s recovering 2026-03-09T17:29:34.765 INFO:teuthology.orchestra.run.vm11.stderr:+ systemctl stop nfs-server 2026-03-09T17:29:34.771 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T17:29:34.775 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm10.local 2026-03-09T17:29:34.775 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'ceph fs volume create foofs' 2026-03-09T17:29:36.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:35 vm10 ceph-mon[53712]: from='client.14502 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:36.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:35 vm10 ceph-mon[53712]: from='client.14506 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "foofs", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:36.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:35 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]: dispatch 2026-03-09T17:29:36.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:35 vm11 ceph-mon[53973]: from='client.14502 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:36.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:35 vm11 ceph-mon[53973]: from='client.14506 -' entity='client.admin' cmd=[{"prefix": "fs volume create", "name": "foofs", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:36.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:35 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]: dispatch 2026-03-09T17:29:37.144 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:36 vm10 ceph-01c7db6a-1bdd-11f1-ac39-b134231210ea-mon-vm10[53707]: 2026-03-09T17:29:36.829+0000 7fb4ebd42700 -1 log_channel(cluster) log [ERR] : Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-09T17:29:37.144 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]': finished 2026-03-09T17:29:37.295 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-09T17:29:37.298 INFO:tasks.cephadm:Waiting for ceph service mds.foofs to start (timeout 300)... 2026-03-09T17:29:37.298 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch ls -f json 2026-03-09T17:29:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: osdmap e45: 8 total, 8 up, 8 in 2026-03-09T17:29:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]: dispatch 2026-03-09T17:29:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: pgmap v97: 33 pgs: 32 unknown, 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:29:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]': finished 2026-03-09T17:29:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: osdmap e46: 8 total, 8 up, 8 in 2026-03-09T17:29:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]: dispatch 2026-03-09T17:29:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-09T17:29:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:37 vm10 ceph-mon[53712]: Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.meta"}]': finished 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: osdmap e45: 8 total, 8 up, 8 in 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]: dispatch 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: pgmap v97: 33 pgs: 32 unknown, 1 active+clean; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool create", "pool": "cephfs.foofs.data"}]': finished 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: osdmap e46: 8 total, 8 up, 8 in 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]: dispatch 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: Health check failed: 1 filesystem is offline (MDS_ALL_DOWN) 2026-03-09T17:29:37.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:37 vm11 ceph-mon[53973]: Health check failed: 1 filesystem is online with fewer MDS than max_mds (MDS_UP_LESS_THAN_MAX) 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: osdmap e47: 8 total, 8 up, 8 in 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]': finished 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: fsmap foofs:0 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: Saving service mds.foofs spec with placement count:2 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm10.ditqnf", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm10.ditqnf", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:38.151 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:38 vm10 ceph-mon[53712]: osdmap e48: 8 total, 8 up, 8 in 2026-03-09T17:29:38.268 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:38.269 INFO:teuthology.orchestra.run.vm10.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-09T17:26:37.139587Z", "last_refresh": "2026-03-09T17:28:34.065299Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-09T17:26:35.316433Z", "last_refresh": "2026-03-09T17:28:34.065329Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-09T17:26:36.142067Z", "last_refresh": "2026-03-09T17:28:34.065357Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-09T17:29:37.234093Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-09T17:29:37.043550Z", "running": 0, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-09T17:26:34.948440Z", "last_refresh": "2026-03-09T17:28:34.065262Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm10:192.168.123.110=vm10", "vm11:192.168.123.111=vm11"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-09T17:26:54.591540Z", "last_refresh": "2026-03-09T17:28:34.065154Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-09T17:26:36.708920Z", "last_refresh": "2026-03-09T17:28:34.065386Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-09T17:28:34.065454Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-09T17:26:35.628870Z", "last_refresh": "2026-03-09T17:28:34.065427Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-09T17:29:38.454 INFO:tasks.cephadm:mds.foofs has 0/2 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: osdmap e47: 8 total, 8 up, 8 in 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "fs new", "fs_name": "foofs", "metadata": "cephfs.foofs.meta", "data": "cephfs.foofs.data"}]': finished 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: fsmap foofs:0 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: Saving service mds.foofs spec with placement count:2 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm10.ditqnf", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm10.ditqnf", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:38.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:38 vm11 ceph-mon[53973]: osdmap e48: 8 total, 8 up, 8 in 2026-03-09T17:29:39.362 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:39 vm11 ceph-mon[53973]: Deploying daemon mds.foofs.vm10.ditqnf on vm10 2026-03-09T17:29:39.362 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:39 vm11 ceph-mon[53973]: pgmap v101: 65 pgs: 14 active+clean, 23 creating+peering, 28 unknown; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:29:39.362 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:39 vm11 ceph-mon[53973]: from='client.14510 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:39.362 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:39 vm11 ceph-mon[53973]: osdmap e49: 8 total, 8 up, 8 in 2026-03-09T17:29:39.362 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:39.362 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm11.sarusb", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-09T17:29:39.362 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm11.sarusb", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-09T17:29:39.362 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:39 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:39 vm10 ceph-mon[53712]: Deploying daemon mds.foofs.vm10.ditqnf on vm10 2026-03-09T17:29:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:39 vm10 ceph-mon[53712]: pgmap v101: 65 pgs: 14 active+clean, 23 creating+peering, 28 unknown; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:29:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:39 vm10 ceph-mon[53712]: from='client.14510 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:39 vm10 ceph-mon[53712]: osdmap e49: 8 total, 8 up, 8 in 2026-03-09T17:29:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm11.sarusb", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]: dispatch 2026-03-09T17:29:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "mds.foofs.vm11.sarusb", "caps": ["mon", "profile mds", "osd", "allow rw tag cephfs *=*", "mds", "allow"]}]': finished 2026-03-09T17:29:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:39 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:39.454 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch ls -f json 2026-03-09T17:29:40.153 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:40.154 INFO:teuthology.orchestra.run.vm10.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-09T17:26:37.139587Z", "last_refresh": "2026-03-09T17:28:34.065299Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-09T17:26:35.316433Z", "last_refresh": "2026-03-09T17:28:34.065329Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-09T17:26:36.142067Z", "last_refresh": "2026-03-09T17:28:34.065357Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-09T17:29:37.234093Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-09T17:29:37.043550Z", "running": 0, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-09T17:26:34.948440Z", "last_refresh": "2026-03-09T17:28:34.065262Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm10:192.168.123.110=vm10", "vm11:192.168.123.111=vm11"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-09T17:26:54.591540Z", "last_refresh": "2026-03-09T17:28:34.065154Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-09T17:26:36.708920Z", "last_refresh": "2026-03-09T17:28:34.065386Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-09T17:28:34.065454Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-09T17:26:35.628870Z", "last_refresh": "2026-03-09T17:28:34.065427Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-09T17:29:40.236 INFO:tasks.cephadm:mds.foofs has 0/2 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: Deploying daemon mds.foofs.vm11.sarusb on vm11 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: mds.? [v2:192.168.123.110:6834/3481240628,v1:192.168.123.110:6835/3481240628] up:boot 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: daemon mds.foofs.vm10.ditqnf assigned to filesystem foofs as rank 0 (now has 1 ranks) 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: Cluster is now healthy 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: fsmap foofs:0 1 up:standby 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mds metadata", "who": "foofs.vm10.ditqnf"}]: dispatch 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: fsmap foofs:1 {0=foofs.vm10.ditqnf=up:creating} 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:40.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:40 vm11 ceph-mon[53973]: daemon mds.foofs.vm10.ditqnf is now active in filesystem foofs as rank 0 2026-03-09T17:29:40.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: Deploying daemon mds.foofs.vm11.sarusb on vm11 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: mds.? [v2:192.168.123.110:6834/3481240628,v1:192.168.123.110:6835/3481240628] up:boot 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: daemon mds.foofs.vm10.ditqnf assigned to filesystem foofs as rank 0 (now has 1 ranks) 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: Health check cleared: MDS_ALL_DOWN (was: 1 filesystem is offline) 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: Health check cleared: MDS_UP_LESS_THAN_MAX (was: 1 filesystem is online with fewer MDS than max_mds) 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: Cluster is now healthy 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: fsmap foofs:0 1 up:standby 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mds metadata", "who": "foofs.vm10.ditqnf"}]: dispatch 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: fsmap foofs:1 {0=foofs.vm10.ditqnf=up:creating} 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:40.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:40 vm10 ceph-mon[53712]: daemon mds.foofs.vm10.ditqnf is now active in filesystem foofs as rank 0 2026-03-09T17:29:41.237 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch ls -f json 2026-03-09T17:29:41.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:41 vm11 ceph-mon[53973]: pgmap v103: 65 pgs: 31 active+clean, 23 creating+peering, 11 unknown; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:29:41.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:41 vm11 ceph-mon[53973]: from='client.14518 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:41.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:41 vm11 ceph-mon[53973]: mds.? [v2:192.168.123.111:6832/1902118686,v1:192.168.123.111:6833/1902118686] up:boot 2026-03-09T17:29:41.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:41 vm11 ceph-mon[53973]: mds.? [v2:192.168.123.110:6834/3481240628,v1:192.168.123.110:6835/3481240628] up:active 2026-03-09T17:29:41.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:41 vm11 ceph-mon[53973]: fsmap foofs:1 {0=foofs.vm10.ditqnf=up:active} 1 up:standby 2026-03-09T17:29:41.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:41 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mds metadata", "who": "foofs.vm11.sarusb"}]: dispatch 2026-03-09T17:29:41.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:41 vm11 ceph-mon[53973]: fsmap foofs:1 {0=foofs.vm10.ditqnf=up:active} 1 up:standby 2026-03-09T17:29:41.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:41 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:41.531 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:41 vm10 ceph-mon[53712]: pgmap v103: 65 pgs: 31 active+clean, 23 creating+peering, 11 unknown; 449 KiB data, 48 MiB used, 160 GiB / 160 GiB avail 2026-03-09T17:29:41.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:41 vm10 ceph-mon[53712]: from='client.14518 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:41.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:41 vm10 ceph-mon[53712]: mds.? [v2:192.168.123.111:6832/1902118686,v1:192.168.123.111:6833/1902118686] up:boot 2026-03-09T17:29:41.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:41 vm10 ceph-mon[53712]: mds.? [v2:192.168.123.110:6834/3481240628,v1:192.168.123.110:6835/3481240628] up:active 2026-03-09T17:29:41.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:41 vm10 ceph-mon[53712]: fsmap foofs:1 {0=foofs.vm10.ditqnf=up:active} 1 up:standby 2026-03-09T17:29:41.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:41 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "mds metadata", "who": "foofs.vm11.sarusb"}]: dispatch 2026-03-09T17:29:41.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:41 vm10 ceph-mon[53712]: fsmap foofs:1 {0=foofs.vm10.ditqnf=up:active} 1 up:standby 2026-03-09T17:29:41.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:41 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:42.204 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:42.245 INFO:teuthology.orchestra.run.vm10.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-09T17:26:37.139587Z", "last_refresh": "2026-03-09T17:28:34.065299Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-09T17:26:35.316433Z", "last_refresh": "2026-03-09T17:28:34.065329Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-09T17:26:36.142067Z", "last_refresh": "2026-03-09T17:28:34.065357Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-09T17:29:37.234093Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-09T17:29:37.043550Z", "last_refresh": "2026-03-09T17:29:41.065930Z", "running": 1, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-09T17:26:34.948440Z", "last_refresh": "2026-03-09T17:28:34.065262Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm10:192.168.123.110=vm10", "vm11:192.168.123.111=vm11"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-09T17:26:54.591540Z", "last_refresh": "2026-03-09T17:28:34.065154Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-09T17:26:36.708920Z", "last_refresh": "2026-03-09T17:28:34.065386Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-09T17:28:34.065454Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-09T17:26:35.628870Z", "last_refresh": "2026-03-09T17:28:34.065427Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-09T17:29:42.689 INFO:tasks.cephadm:mds.foofs has 1/2 2026-03-09T17:29:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:43 vm10 ceph-mon[53712]: pgmap v104: 65 pgs: 59 active+clean, 6 creating+peering; 450 KiB data, 49 MiB used, 160 GiB / 160 GiB avail; 784 B/s wr, 3 op/s 2026-03-09T17:29:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:43 vm10 ceph-mon[53712]: from='client.14522 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:43 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:43 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:43 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:43 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:43 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:43 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:43.690 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- ceph orch ls -f json 2026-03-09T17:29:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:43 vm11 ceph-mon[53973]: pgmap v104: 65 pgs: 59 active+clean, 6 creating+peering; 450 KiB data, 49 MiB used, 160 GiB / 160 GiB avail; 784 B/s wr, 3 op/s 2026-03-09T17:29:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:43 vm11 ceph-mon[53973]: from='client.14522 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:43 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:43 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:43 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:43 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:43 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:43 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:44.429 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T17:29:44.429 INFO:teuthology.orchestra.run.vm10.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-09T17:26:37.139587Z", "last_refresh": "2026-03-09T17:29:42.220592Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-09T17:26:35.316433Z", "last_refresh": "2026-03-09T17:29:41.065447Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "status": {"created": "2026-03-09T17:26:36.142067Z", "last_refresh": "2026-03-09T17:29:42.220660Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-09T17:29:37.234093Z service:mds.foofs [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foofs", "service_name": "mds.foofs", "service_type": "mds", "status": {"created": "2026-03-09T17:29:37.043550Z", "last_refresh": "2026-03-09T17:29:41.065930Z", "running": 2, "size": 2}}, {"placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-09T17:26:34.948440Z", "last_refresh": "2026-03-09T17:29:41.065608Z", "running": 2, "size": 2}}, {"placement": {"count": 2, "hosts": ["vm10:192.168.123.110=vm10", "vm11:192.168.123.111=vm11"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-09T17:26:54.591540Z", "last_refresh": "2026-03-09T17:29:41.065660Z", "running": 2, "size": 2}}, {"placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-09T17:26:36.708920Z", "last_refresh": "2026-03-09T17:29:41.065719Z", "ports": [9100], "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "e1d6a67b021eb077ee22bf650f1a9fb1980a2cf5c36bdb9cba9eac6de8f702d9", "container_image_name": "quay.io/ceph/ceph@sha256:12a0a4f43413fd97a14a3d47a3451b2d2df50020835bb93db666209f3f77617a", "last_refresh": "2026-03-09T17:29:41.065765Z", "running": 8, "size": 8}, "unmanaged": true}, {"placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-09T17:26:35.628870Z", "last_refresh": "2026-03-09T17:29:42.220723Z", "ports": [9095], "running": 1, "size": 1}}] 2026-03-09T17:29:44.522 INFO:tasks.cephadm:mds.foofs has 2/2 2026-03-09T17:29:44.522 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T17:29:44.527 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm10.local 2026-03-09T17:29:44.527 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'ceph nfs cluster create foo --placement=2 || ceph nfs cluster create cephfs foo --placement=2' 2026-03-09T17:29:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:45 vm11 ceph-mon[53973]: pgmap v105: 65 pgs: 65 active+clean; 450 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 871 B/s wr, 5 op/s 2026-03-09T17:29:45.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:45.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:45 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool create", "pool": ".nfs"}]: dispatch 2026-03-09T17:29:45.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:45 vm10 ceph-mon[53712]: pgmap v105: 65 pgs: 65 active+clean; 450 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 871 B/s wr, 5 op/s 2026-03-09T17:29:45.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:45.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:45 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool create", "pool": ".nfs"}]: dispatch 2026-03-09T17:29:46.685 INFO:teuthology.orchestra.run.vm10.stdout:NFS Cluster Created Successfully 2026-03-09T17:29:46.729 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:46 vm10 ceph-mon[53712]: from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:46.730 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:46 vm10 ceph-mon[53712]: from='client.14530 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "foo", "placement": "2", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:46.730 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:46 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool create", "pool": ".nfs"}]': finished 2026-03-09T17:29:46.730 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:46 vm10 ceph-mon[53712]: osdmap e50: 8 total, 8 up, 8 in 2026-03-09T17:29:46.730 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:46 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-09T17:29:46.730 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:46 vm10 ceph-mon[53712]: pgmap v107: 97 pgs: 32 unknown, 65 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s wr, 5 op/s 2026-03-09T17:29:46.745 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'ceph nfs export create cephfs --fsname foofs --clusterid foo --binding /fake || ceph nfs export create cephfs --fsname foofs --cluster-id foo --pseudo-path /fake' 2026-03-09T17:29:46.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:46 vm11 ceph-mon[53973]: from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T17:29:46.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:46 vm11 ceph-mon[53973]: from='client.14530 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "foo", "placement": "2", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:46.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:46 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool create", "pool": ".nfs"}]': finished 2026-03-09T17:29:46.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:46 vm11 ceph-mon[53973]: osdmap e50: 8 total, 8 up, 8 in 2026-03-09T17:29:46.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:46 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-09T17:29:46.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:46 vm11 ceph-mon[53973]: pgmap v107: 97 pgs: 32 unknown, 65 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 1.6 KiB/s wr, 5 op/s 2026-03-09T17:29:47.366 INFO:teuthology.orchestra.run.vm10.stderr:Invalid command: Unexpected argument '--clusterid' 2026-03-09T17:29:47.366 INFO:teuthology.orchestra.run.vm10.stderr:nfs export create cephfs [] [--readonly] [--client_addr ...] [--squash ] : Create a CephFS export 2026-03-09T17:29:47.367 INFO:teuthology.orchestra.run.vm10.stderr:Error EINVAL: invalid command 2026-03-09T17:29:47.653 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-09T17:29:47.653 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: osdmap e51: 8 total, 8 up, 8 in 2026-03-09T17:29:47.653 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: Saving service nfs.foo spec with placement count:2 2026-03-09T17:29:47.653 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:47.653 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:47.653 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:47.653 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:47.653 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:47.710 INFO:teuthology.orchestra.run.vm10.stdout:{ 2026-03-09T17:29:47.711 INFO:teuthology.orchestra.run.vm10.stdout: "bind": "/fake", 2026-03-09T17:29:47.711 INFO:teuthology.orchestra.run.vm10.stdout: "fs": "foofs", 2026-03-09T17:29:47.711 INFO:teuthology.orchestra.run.vm10.stdout: "path": "/", 2026-03-09T17:29:47.711 INFO:teuthology.orchestra.run.vm10.stdout: "cluster": "foo", 2026-03-09T17:29:47.711 INFO:teuthology.orchestra.run.vm10.stdout: "mode": "RW" 2026-03-09T17:29:47.711 INFO:teuthology.orchestra.run.vm10.stdout:} 2026-03-09T17:29:47.814 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v17.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 01c7db6a-1bdd-11f1-ac39-b134231210ea -- bash -c 'while ! ceph orch ls | grep nfs | grep 2/2 ; do sleep 1 ; done' 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: Creating key for client.nfs.foo.0.0.vm10.fkkjeb 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm10.fkkjeb", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm10.fkkjeb", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: Ensuring nfs.foo.0 is in the ganesha grace table 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: Rados config object exists: conf-nfs.foo 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: Creating key for client.nfs.foo.0.0.vm10.fkkjeb-rgw 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm10.fkkjeb-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm10.fkkjeb-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T17:29:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:47 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: osdmap e51: 8 total, 8 up, 8 in 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: Saving service nfs.foo spec with placement count:2 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: Creating key for client.nfs.foo.0.0.vm10.fkkjeb 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm10.fkkjeb", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm10.fkkjeb", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: Ensuring nfs.foo.0 is in the ganesha grace table 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: Rados config object exists: conf-nfs.foo 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: Creating key for client.nfs.foo.0.0.vm10.fkkjeb-rgw 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm10.fkkjeb-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm10.fkkjeb-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T17:29:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:47 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:48.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:48 vm10 ceph-mon[53712]: Deploying daemon nfs.foo.0.0.vm10.fkkjeb on vm10 2026-03-09T17:29:48.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:48 vm10 ceph-mon[53712]: from='client.14550 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "foofs", "cluster_id": "foo", "pseudo_path": "/fake", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:48.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:48 vm10 ceph-mon[53712]: osdmap e52: 8 total, 8 up, 8 in 2026-03-09T17:29:48.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:48 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-09T17:29:48.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:48 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-09T17:29:48.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:48 vm10 ceph-mon[53712]: pgmap v110: 97 pgs: 14 creating+peering, 10 unknown, 73 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 2.0 KiB/s wr, 5 op/s 2026-03-09T17:29:48.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:48 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:49.016 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:48 vm11 ceph-mon[53973]: Deploying daemon nfs.foo.0.0.vm10.fkkjeb on vm10 2026-03-09T17:29:49.016 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:48 vm11 ceph-mon[53973]: from='client.14550 -' entity='client.admin' cmd=[{"prefix": "nfs export create cephfs", "fsname": "foofs", "cluster_id": "foo", "pseudo_path": "/fake", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:49.016 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:48 vm11 ceph-mon[53973]: osdmap e52: 8 total, 8 up, 8 in 2026-03-09T17:29:49.016 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:48 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]: dispatch 2026-03-09T17:29:49.016 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:48 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo, allow rw tag cephfs data=foofs", "mds", "allow rw path=/"], "format": "json"}]': finished 2026-03-09T17:29:49.016 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:48 vm11 ceph-mon[53973]: pgmap v110: 97 pgs: 14 creating+peering, 10 unknown, 73 active+clean; 451 KiB data, 50 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 2.0 KiB/s wr, 5 op/s 2026-03-09T17:29:49.016 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:48 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: Creating key for client.nfs.foo.1.0.vm11.bwajpk 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm11.bwajpk", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm11.bwajpk", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: Ensuring nfs.foo.1 is in the ganesha grace table 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: mgrmap e20: vm10.qompzp(active, since 2m), standbys: vm11.ohhvcs 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='client.14554 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-09T17:29:49.824 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-09T17:29:49.825 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: Rados config object exists: conf-nfs.foo 2026-03-09T17:29:49.825 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: Creating key for client.nfs.foo.1.0.vm11.bwajpk-rgw 2026-03-09T17:29:49.825 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm11.bwajpk-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T17:29:49.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm11.bwajpk-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T17:29:49.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:49.832 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:49 vm11 ceph-mon[53973]: Deploying daemon nfs.foo.1.0.vm11.bwajpk on vm11 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: Creating key for client.nfs.foo.1.0.vm11.bwajpk 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm11.bwajpk", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm11.bwajpk", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: Ensuring nfs.foo.1 is in the ganesha grace table 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: mgrmap e20: vm10.qompzp(active, since 2m), standbys: vm11.ohhvcs 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='client.14554 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: Rados config object exists: conf-nfs.foo 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: Creating key for client.nfs.foo.1.0.vm11.bwajpk-rgw 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm11.bwajpk-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.1.0.vm11.bwajpk-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:50.186 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:49 vm10 ceph-mon[53712]: Deploying daemon nfs.foo.1.0.vm11.bwajpk on vm11 2026-03-09T17:29:50.824 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:50 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:50.824 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:50 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:50.824 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:50 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:50.824 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:50 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:50.824 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:50 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:50.824 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:50 vm10 ceph-mon[53712]: from='client.14574 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:50.824 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:50 vm10 ceph-mon[53712]: pgmap v111: 97 pgs: 14 creating+peering, 83 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 1.7 KiB/s wr, 2 op/s 2026-03-09T17:29:50.825 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:50 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:50.825 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:50 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:50.825 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:50 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:50.825 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:50 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:50.825 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:50 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:50.826 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:50 vm11 ceph-mon[53973]: from='client.14574 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:50.826 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:50 vm11 ceph-mon[53973]: pgmap v111: 97 pgs: 14 creating+peering, 83 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 1.7 KiB/s wr, 2 op/s 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='client.14584 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.259 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='client.14584 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:52.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:29:52.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:29:52.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:29:52.267 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.267 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:52.498 INFO:teuthology.orchestra.run.vm10.stdout:nfs.foo ?:2049 2/2 1s ago 5s count:2 2026-03-09T17:29:52.561 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-09T17:29:52.564 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm10.local 2026-03-09T17:29:52.564 DEBUG:teuthology.orchestra.run.vm10:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'mkdir /mnt/foo' 2026-03-09T17:29:52.594 INFO:teuthology.orchestra.run.vm10.stderr:+ mkdir /mnt/foo 2026-03-09T17:29:52.596 DEBUG:teuthology.orchestra.run.vm10:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'while ! mount -t nfs $(hostname):/fake /mnt/foo -o sync ; do sleep 5 ; done' 2026-03-09T17:29:52.666 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:29:52.666 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:29:52.961 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:29:52.962 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:29:53.515 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:53 vm10 ceph-mon[53712]: pgmap v112: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 477 B/s rd, 636 B/s wr, 1 op/s 2026-03-09T17:29:53.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:53 vm11 ceph-mon[53973]: pgmap v112: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 477 B/s rd, 636 B/s wr, 1 op/s 2026-03-09T17:29:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:54 vm11 ceph-mon[53973]: from='client.24389 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:54 vm10 ceph-mon[53712]: from='client.24389 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T17:29:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:55 vm11 ceph-mon[53973]: pgmap v113: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 4.6 KiB/s rd, 2.4 KiB/s wr, 7 op/s 2026-03-09T17:29:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:55 vm10 ceph-mon[53712]: pgmap v113: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 4.6 KiB/s rd, 2.4 KiB/s wr, 7 op/s 2026-03-09T17:29:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:29:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:57 vm11 ceph-mon[53973]: pgmap v114: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 3.9 KiB/s rd, 2.0 KiB/s wr, 6 op/s 2026-03-09T17:29:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:57 vm10 ceph-mon[53712]: pgmap v114: 97 pgs: 97 active+clean; 451 KiB data, 51 MiB used, 160 GiB / 160 GiB avail; 3.9 KiB/s rd, 2.0 KiB/s wr, 6 op/s 2026-03-09T17:29:57.963 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:29:57.964 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:29:57.990 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:29:57.991 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:29:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:29:59 vm11 ceph-mon[53973]: pgmap v115: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.6 KiB/s rd, 2.6 KiB/s wr, 7 op/s 2026-03-09T17:29:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:29:59 vm10 ceph-mon[53712]: pgmap v115: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.6 KiB/s rd, 2.6 KiB/s wr, 7 op/s 2026-03-09T17:30:00.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:00 vm11 ceph-mon[53973]: overall HEALTH_OK 2026-03-09T17:30:00.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:00 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:30:00.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:00 vm10 ceph-mon[53712]: overall HEALTH_OK 2026-03-09T17:30:00.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:00 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:30:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:01 vm11 ceph-mon[53973]: pgmap v116: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.0 KiB/s rd, 2.0 KiB/s wr, 5 op/s 2026-03-09T17:30:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:01 vm10 ceph-mon[53712]: pgmap v116: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.0 KiB/s rd, 2.0 KiB/s wr, 5 op/s 2026-03-09T17:30:02.992 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:02.993 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:03.020 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:03.020 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:03 vm11 ceph-mon[53973]: pgmap v117: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.0 KiB/s rd, 2.1 KiB/s wr, 5 op/s 2026-03-09T17:30:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:03 vm10 ceph-mon[53712]: pgmap v117: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.0 KiB/s rd, 2.1 KiB/s wr, 5 op/s 2026-03-09T17:30:04.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:04 vm10 ceph-mon[53712]: pgmap v118: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.0 KiB/s rd, 2.0 KiB/s wr, 5 op/s 2026-03-09T17:30:04.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:04 vm11 ceph-mon[53973]: pgmap v118: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 3.0 KiB/s rd, 2.0 KiB/s wr, 5 op/s 2026-03-09T17:30:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:07 vm10 ceph-mon[53712]: pgmap v119: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 767 B/s wr, 1 op/s 2026-03-09T17:30:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:07 vm11 ceph-mon[53973]: pgmap v119: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 767 B/s wr, 1 op/s 2026-03-09T17:30:08.022 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:08.022 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:08.067 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:08.067 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:09 vm11 ceph-mon[53973]: pgmap v120: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 853 B/s wr, 1 op/s 2026-03-09T17:30:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:09 vm10 ceph-mon[53712]: pgmap v120: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 853 B/s wr, 1 op/s 2026-03-09T17:30:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:11 vm11 ceph-mon[53973]: pgmap v121: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:11.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:11 vm10 ceph-mon[53712]: pgmap v121: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:13.069 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:13.069 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:13.120 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:13.120 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:13.516 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:13 vm10 ceph-mon[53712]: pgmap v122: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:13.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:13 vm11 ceph-mon[53973]: pgmap v122: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:15 vm11 ceph-mon[53973]: pgmap v123: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:15 vm10 ceph-mon[53712]: pgmap v123: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:17 vm11 ceph-mon[53973]: pgmap v124: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:17 vm10 ceph-mon[53712]: pgmap v124: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:18.122 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:18.122 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:18.185 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:18.186 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:19 vm11 ceph-mon[53973]: pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:19 vm10 ceph-mon[53712]: pgmap v125: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:21.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:21 vm11 ceph-mon[53973]: pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:21.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:21 vm10 ceph-mon[53712]: pgmap v126: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:23.187 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:23.188 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:23.217 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:23.217 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:23 vm11 ceph-mon[53973]: pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:23.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:23 vm10 ceph-mon[53712]: pgmap v127: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:25 vm11 ceph-mon[53973]: pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:25 vm10 ceph-mon[53712]: pgmap v128: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:27.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:27 vm10 ceph-mon[53712]: pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:27 vm11 ceph-mon[53973]: pgmap v129: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:28.219 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:28.219 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:28.248 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:28.248 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.18", "id": [7, 2]}]: dispatch 2026-03-09T17:30:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-09T17:30:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [1, 2]}]: dispatch 2026-03-09T17:30:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1c", "id": [1, 5]}]: dispatch 2026-03-09T17:30:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:28 vm11 ceph-mon[53973]: pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:30:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:30:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.18", "id": [7, 2]}]: dispatch 2026-03-09T17:30:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]: dispatch 2026-03-09T17:30:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [1, 2]}]: dispatch 2026-03-09T17:30:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1c", "id": [1, 5]}]: dispatch 2026-03-09T17:30:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:28 vm10 ceph-mon[53712]: pgmap v130: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:30:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:30:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:29 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.18", "id": [7, 2]}]': finished 2026-03-09T17:30:29.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:29 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]': finished 2026-03-09T17:30:29.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:29 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [1, 2]}]': finished 2026-03-09T17:30:29.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:29 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1c", "id": [1, 5]}]': finished 2026-03-09T17:30:29.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:29 vm11 ceph-mon[53973]: osdmap e53: 8 total, 8 up, 8 in 2026-03-09T17:30:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.18", "id": [7, 2]}]': finished 2026-03-09T17:30:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.2", "id": [1, 2]}]': finished 2026-03-09T17:30:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.10", "id": [1, 2]}]': finished 2026-03-09T17:30:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:29 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.1c", "id": [1, 5]}]': finished 2026-03-09T17:30:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:29 vm10 ceph-mon[53712]: osdmap e53: 8 total, 8 up, 8 in 2026-03-09T17:30:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:30 vm11 ceph-mon[53973]: osdmap e54: 8 total, 8 up, 8 in 2026-03-09T17:30:30.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:30 vm11 ceph-mon[53973]: pgmap v133: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 127 B/s wr, 0 op/s 2026-03-09T17:30:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:30 vm10 ceph-mon[53712]: osdmap e54: 8 total, 8 up, 8 in 2026-03-09T17:30:30.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:30 vm10 ceph-mon[53712]: pgmap v133: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 127 B/s wr, 0 op/s 2026-03-09T17:30:33.250 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:33.250 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:33.314 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:33.314 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:33 vm10 ceph-mon[53712]: pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s wr, 0 op/s 2026-03-09T17:30:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:33 vm11 ceph-mon[53973]: pgmap v134: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s wr, 0 op/s 2026-03-09T17:30:34.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:34 vm10 ceph-mon[53712]: pgmap v135: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:34.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:34 vm11 ceph-mon[53973]: pgmap v135: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:37 vm10 ceph-mon[53712]: pgmap v136: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 127 B/s wr, 0 op/s 2026-03-09T17:30:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:37 vm11 ceph-mon[53973]: pgmap v136: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 127 B/s wr, 0 op/s 2026-03-09T17:30:38.316 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:38.316 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:38.374 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:38.374 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:39.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:39 vm10 ceph-mon[53712]: pgmap v137: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 213 B/s rd, 213 B/s wr, 0 op/s 2026-03-09T17:30:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:39 vm11 ceph-mon[53973]: pgmap v137: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 213 B/s rd, 213 B/s wr, 0 op/s 2026-03-09T17:30:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:41 vm10 ceph-mon[53712]: pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 193 B/s rd, 193 B/s wr, 0 op/s 2026-03-09T17:30:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:41 vm11 ceph-mon[53973]: pgmap v138: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 193 B/s rd, 193 B/s wr, 0 op/s 2026-03-09T17:30:43.376 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:43.376 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:43.405 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:43.405 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:43 vm10 ceph-mon[53712]: pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:43 vm11 ceph-mon[53973]: pgmap v139: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:44.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:44 vm10 ceph-mon[53712]: pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:44.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:44 vm11 ceph-mon[53973]: pgmap v140: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:47 vm10 ceph-mon[53712]: pgmap v141: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:47 vm11 ceph-mon[53973]: pgmap v141: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:48.407 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:48.407 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:48.432 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:48.433 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:49 vm11 ceph-mon[53973]: pgmap v142: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:49.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:49 vm10 ceph-mon[53712]: pgmap v142: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:51 vm11 ceph-mon[53973]: pgmap v143: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:51 vm10 ceph-mon[53712]: pgmap v143: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:30:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:30:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:30:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:30:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:30:52.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:30:52.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:30:52.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:30:52.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:30:52.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:30:53.434 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:53.435 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:53.461 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:53.461 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:53 vm11 ceph-mon[53973]: pgmap v144: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:53 vm10 ceph-mon[53712]: pgmap v144: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:55 vm11 ceph-mon[53973]: pgmap v145: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:55 vm10 ceph-mon[53712]: pgmap v145: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:57 vm11 ceph-mon[53973]: pgmap v146: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:57 vm10 ceph-mon[53712]: pgmap v146: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:30:58.463 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:30:58.463 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:30:58.488 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:30:58.489 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:30:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:30:59 vm10 ceph-mon[53712]: pgmap v147: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:30:59.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:30:59 vm11 ceph-mon[53973]: pgmap v147: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:00.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:00 vm10 ceph-mon[53712]: pgmap v148: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:00.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:00 vm11 ceph-mon[53973]: pgmap v148: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:03.490 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:03.490 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:03.515 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:03.515 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:03 vm10 ceph-mon[53712]: pgmap v149: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:03 vm11 ceph-mon[53973]: pgmap v149: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:05 vm10 ceph-mon[53712]: pgmap v150: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:05 vm11 ceph-mon[53973]: pgmap v150: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:07.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:07 vm11 ceph-mon[53973]: pgmap v151: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:07 vm10 ceph-mon[53712]: pgmap v151: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:08.517 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:08.517 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:08.543 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:08.543 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:08.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:08 vm10 ceph-mon[53712]: pgmap v152: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:08 vm11 ceph-mon[53973]: pgmap v152: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:11 vm10 ceph-mon[53712]: pgmap v153: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:11 vm11 ceph-mon[53973]: pgmap v153: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:13.545 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:13.545 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:13.571 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:13.572 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:13 vm11 ceph-mon[53973]: pgmap v154: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:13 vm10 ceph-mon[53712]: pgmap v154: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:15 vm11 ceph-mon[53973]: pgmap v155: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:15 vm10 ceph-mon[53712]: pgmap v155: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:17 vm11 ceph-mon[53973]: pgmap v156: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:17 vm10 ceph-mon[53712]: pgmap v156: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:18.573 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:18.574 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:18.600 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:18.601 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:19 vm10 ceph-mon[53712]: pgmap v157: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:19.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:19 vm11 ceph-mon[53973]: pgmap v157: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:20.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:20 vm10 ceph-mon[53712]: pgmap v158: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:20.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:20 vm11 ceph-mon[53973]: pgmap v158: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:23 vm11 ceph-mon[53973]: pgmap v159: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:23.603 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:23.603 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:23.630 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:23.630 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:23.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:23 vm10 ceph-mon[53712]: pgmap v159: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:25 vm11 ceph-mon[53973]: pgmap v160: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:25.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:25 vm10 ceph-mon[53712]: pgmap v160: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:27.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:27 vm11 ceph-mon[53973]: pgmap v161: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:27 vm10 ceph-mon[53712]: pgmap v161: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:28.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:31:28.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:31:28.632 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:28.632 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:28.658 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:28.659 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:28.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:31:28.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:31:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:29 vm11 ceph-mon[53973]: pgmap v162: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:29 vm10 ceph-mon[53712]: pgmap v162: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:31 vm11 ceph-mon[53973]: pgmap v163: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:31 vm10 ceph-mon[53712]: pgmap v163: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:33 vm11 ceph-mon[53973]: pgmap v164: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:33.660 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:33.661 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:33.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:33 vm10 ceph-mon[53712]: pgmap v164: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:33.686 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:33.687 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:35 vm10 ceph-mon[53712]: pgmap v165: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:35.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:35 vm11 ceph-mon[53973]: pgmap v165: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:37.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:37 vm11 ceph-mon[53973]: pgmap v166: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:37 vm10 ceph-mon[53712]: pgmap v166: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:38.688 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:38.689 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:38.715 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:38.715 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:39.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:39 vm10 ceph-mon[53712]: pgmap v167: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:39 vm11 ceph-mon[53973]: pgmap v167: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:41 vm10 ceph-mon[53712]: pgmap v168: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:41 vm11 ceph-mon[53973]: pgmap v168: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:43 vm10 ceph-mon[53712]: pgmap v169: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:43.717 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:43.717 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:43.743 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:43.744 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:43 vm11 ceph-mon[53973]: pgmap v169: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:44.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:44 vm10 ceph-mon[53712]: pgmap v170: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:44.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:44 vm11 ceph-mon[53973]: pgmap v170: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:47.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:47 vm10 ceph-mon[53712]: pgmap v171: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:47 vm11 ceph-mon[53973]: pgmap v171: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:48.745 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:48.746 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:48.782 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:48.782 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:49 vm11 ceph-mon[53973]: pgmap v172: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:49.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:49 vm10 ceph-mon[53712]: pgmap v172: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:51.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:51 vm11 ceph-mon[53973]: pgmap v173: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:51.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:51 vm10 ceph-mon[53712]: pgmap v173: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:31:52.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:31:52.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:31:52.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:31:52.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:31:52.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:31:52.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:31:52.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:31:52.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:31:52.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:31:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:53 vm11 ceph-mon[53973]: pgmap v174: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:53 vm10 ceph-mon[53712]: pgmap v174: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:53.784 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:53.784 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:53.811 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:53.812 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:55 vm10 ceph-mon[53712]: pgmap v175: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:55.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:55 vm11 ceph-mon[53973]: pgmap v175: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:56.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:56 vm10 ceph-mon[53712]: pgmap v176: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:56 vm11 ceph-mon[53973]: pgmap v176: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:31:58.814 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:31:58.814 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:31:58.840 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:31:58.840 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:31:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:31:59 vm11 ceph-mon[53973]: pgmap v177: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:31:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:31:59 vm10 ceph-mon[53712]: pgmap v177: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:01 vm11 ceph-mon[53973]: pgmap v178: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:01.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:01 vm10 ceph-mon[53712]: pgmap v178: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:03.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:03 vm11 ceph-mon[53973]: pgmap v179: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:03 vm10 ceph-mon[53712]: pgmap v179: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:03.842 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:03.842 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:03.868 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:03.869 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:05 vm11 ceph-mon[53973]: pgmap v180: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:05 vm10 ceph-mon[53712]: pgmap v180: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:07.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:07 vm11 ceph-mon[53973]: pgmap v181: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:07 vm10 ceph-mon[53712]: pgmap v181: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:08.870 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:08.871 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:08.896 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:08.896 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:09 vm10 ceph-mon[53712]: pgmap v182: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:09 vm11 ceph-mon[53973]: pgmap v182: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:11 vm10 ceph-mon[53712]: pgmap v183: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:11 vm11 ceph-mon[53973]: pgmap v183: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:13.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:13 vm10 ceph-mon[53712]: pgmap v184: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:13 vm11 ceph-mon[53973]: pgmap v184: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:13.898 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:13.899 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:13.926 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:13.927 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:14.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:14 vm11 ceph-mon[53973]: pgmap v185: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:14.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:14 vm10 ceph-mon[53712]: pgmap v185: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:17.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:17 vm10 ceph-mon[53712]: pgmap v186: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:17 vm11 ceph-mon[53973]: pgmap v186: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:18.928 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:18.929 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:18.954 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:18.955 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:19 vm11 ceph-mon[53973]: pgmap v187: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:19 vm10 ceph-mon[53712]: pgmap v187: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:21 vm11 ceph-mon[53973]: pgmap v188: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:21 vm10 ceph-mon[53712]: pgmap v188: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:23.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:23 vm10 ceph-mon[53712]: pgmap v189: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:23 vm11 ceph-mon[53973]: pgmap v189: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:23.956 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:23.957 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:23.987 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:23.988 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:25 vm10 ceph-mon[53712]: pgmap v190: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:25 vm11 ceph-mon[53973]: pgmap v190: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:27.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:27 vm11 ceph-mon[53973]: pgmap v191: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:27 vm10 ceph-mon[53712]: pgmap v191: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:28.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:32:28.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:32:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:32:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:32:28.989 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:28.990 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:29.021 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:29.021 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:29 vm10 ceph-mon[53712]: pgmap v192: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:29 vm11 ceph-mon[53973]: pgmap v192: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:30 vm10 ceph-mon[53712]: pgmap v193: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:30 vm11 ceph-mon[53973]: pgmap v193: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:33 vm11 ceph-mon[53973]: pgmap v194: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:33.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:33 vm10 ceph-mon[53712]: pgmap v194: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:34.022 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:34.023 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:34.058 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:34.059 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:35.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:35 vm11 ceph-mon[53973]: pgmap v195: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:35 vm10 ceph-mon[53712]: pgmap v195: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:37.664 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:37 vm11 ceph-mon[53973]: pgmap v196: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:37 vm10 ceph-mon[53712]: pgmap v196: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:39.060 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:39.060 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:39.085 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:39.086 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:39.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:39 vm10 ceph-mon[53712]: pgmap v197: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:39 vm11 ceph-mon[53973]: pgmap v197: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:41 vm10 ceph-mon[53712]: pgmap v198: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:41 vm11 ceph-mon[53973]: pgmap v198: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:43 vm10 ceph-mon[53712]: pgmap v199: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:43.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:43 vm11 ceph-mon[53973]: pgmap v199: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:44.087 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:44.088 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:44.113 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:44.114 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:45 vm10 ceph-mon[53712]: pgmap v200: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:45.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:45 vm11 ceph-mon[53973]: pgmap v200: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:46.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:46 vm11 ceph-mon[53973]: pgmap v201: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:46.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:46 vm10 ceph-mon[53712]: pgmap v201: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:49.115 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:49.116 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:49.141 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:49.141 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:49 vm11 ceph-mon[53973]: pgmap v202: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:49.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:49 vm10 ceph-mon[53712]: pgmap v202: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:51 vm10 ceph-mon[53712]: pgmap v203: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:51.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:51 vm11 ceph-mon[53973]: pgmap v203: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:32:52.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:32:52.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:52 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:32:52.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:32:52.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:32:52.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:52 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:32:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:53 vm10 ceph-mon[53712]: pgmap v204: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:32:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:32:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:53 vm11 ceph-mon[53973]: pgmap v204: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:32:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:32:54.143 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:54.143 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:54.169 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:54.170 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:32:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:55 vm10 ceph-mon[53712]: pgmap v205: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:55 vm11 ceph-mon[53973]: pgmap v205: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:57.664 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:57 vm11 ceph-mon[53973]: pgmap v206: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:57 vm10 ceph-mon[53712]: pgmap v206: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:32:58.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:32:58 vm10 ceph-mon[53712]: pgmap v207: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:32:58 vm11 ceph-mon[53973]: pgmap v207: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:32:59.172 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:32:59.172 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:32:59.199 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:32:59.199 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:01 vm10 ceph-mon[53712]: pgmap v208: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:01 vm11 ceph-mon[53973]: pgmap v208: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:03.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:03 vm11 ceph-mon[53973]: pgmap v209: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:03 vm10 ceph-mon[53712]: pgmap v209: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:04.201 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:04.202 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:04.230 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:04.230 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:05 vm10 ceph-mon[53712]: pgmap v210: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:05 vm11 ceph-mon[53973]: pgmap v210: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:07.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:07 vm11 ceph-mon[53973]: pgmap v211: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:07 vm10 ceph-mon[53712]: pgmap v211: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:09.232 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:09.232 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:09.257 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:09.258 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:09 vm10 ceph-mon[53712]: pgmap v212: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:09 vm11 ceph-mon[53973]: pgmap v212: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:11 vm10 ceph-mon[53712]: pgmap v213: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:11 vm11 ceph-mon[53973]: pgmap v213: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:13 vm10 ceph-mon[53712]: pgmap v214: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:13 vm11 ceph-mon[53973]: pgmap v214: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:14.259 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:14.260 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:14.285 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:14.286 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:14.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:14 vm10 ceph-mon[53712]: pgmap v215: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:14.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:14 vm11 ceph-mon[53973]: pgmap v215: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:17 vm10 ceph-mon[53712]: pgmap v216: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:17 vm11 ceph-mon[53973]: pgmap v216: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:19.287 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:19.288 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:19.313 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:19.313 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:19 vm10 ceph-mon[53712]: pgmap v217: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:19.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:19 vm11 ceph-mon[53973]: pgmap v217: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:21 vm10 ceph-mon[53712]: pgmap v218: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:21.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:21 vm11 ceph-mon[53973]: pgmap v218: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:23.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:23 vm10 ceph-mon[53712]: pgmap v219: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:23 vm11 ceph-mon[53973]: pgmap v219: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:24.315 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:24.315 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:24.341 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:24.341 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:25 vm10 ceph-mon[53712]: pgmap v220: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:25 vm11 ceph-mon[53973]: pgmap v220: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:26.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:26 vm10 ceph-mon[53712]: pgmap v221: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:26 vm11 ceph-mon[53973]: pgmap v221: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:28.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:33:28.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:33:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:33:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:33:29.343 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:29.343 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:29.368 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:29.368 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:29 vm10 ceph-mon[53712]: pgmap v222: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:29 vm11 ceph-mon[53973]: pgmap v222: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:31 vm10 ceph-mon[53712]: pgmap v223: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:31 vm11 ceph-mon[53973]: pgmap v223: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:33 vm10 ceph-mon[53712]: pgmap v224: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:33.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:33 vm11 ceph-mon[53973]: pgmap v224: 97 pgs: 97 active+clean; 453 KiB data, 52 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:34.370 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:34.370 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:34.396 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:34.397 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:34.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:34 vm10 ceph-mon[53712]: pgmap v225: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:34.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:34 vm11 ceph-mon[53973]: pgmap v225: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:37.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:37 vm10 ceph-mon[53712]: pgmap v226: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:37 vm11 ceph-mon[53973]: pgmap v226: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:39.398 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:39.398 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:39.528 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:39.550 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:39 vm11 ceph-mon[53973]: pgmap v227: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:39 vm10 ceph-mon[53712]: pgmap v227: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:40 vm11 ceph-mon[53973]: pgmap v228: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:40 vm10 ceph-mon[53712]: pgmap v228: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:43 vm10 ceph-mon[53712]: pgmap v229: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:43 vm11 ceph-mon[53973]: pgmap v229: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:44.530 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:44.531 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:44.555 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:44.556 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:45 vm10 ceph-mon[53712]: pgmap v230: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:45 vm11 ceph-mon[53973]: pgmap v230: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:47.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:47 vm11 ceph-mon[53973]: pgmap v231: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:47 vm10 ceph-mon[53712]: pgmap v231: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:49.557 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:49.558 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:49.583 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:49.584 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:49.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:49 vm10 ceph-mon[53712]: pgmap v232: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:49.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:49 vm11 ceph-mon[53973]: pgmap v232: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:50.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:50 vm11 ceph-mon[53973]: pgmap v233: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:50.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:50 vm10 ceph-mon[53712]: pgmap v233: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:53 vm10 ceph-mon[53712]: pgmap v234: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:33:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:33:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:33:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:33:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:33:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:53 vm11 ceph-mon[53973]: pgmap v234: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:33:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:33:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:33:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:33:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:33:54.585 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:54.586 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:54.614 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:54.615 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:55 vm10 ceph-mon[53712]: pgmap v235: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:55 vm11 ceph-mon[53973]: pgmap v235: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:56 vm11 ceph-mon[53973]: pgmap v236: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:56 vm10 ceph-mon[53712]: pgmap v236: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:33:59.616 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:33:59.617 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:33:59.643 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:33:59.644 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:33:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:33:59 vm10 ceph-mon[53712]: pgmap v237: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:33:59.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:33:59 vm11 ceph-mon[53973]: pgmap v237: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:01 vm10 ceph-mon[53712]: pgmap v238: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:01 vm11 ceph-mon[53973]: pgmap v238: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:03 vm10 ceph-mon[53712]: pgmap v239: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:03 vm11 ceph-mon[53973]: pgmap v239: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:04.645 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:04.645 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:04.671 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:04.672 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:04.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:04 vm10 ceph-mon[53712]: pgmap v240: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:04.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:04 vm11 ceph-mon[53973]: pgmap v240: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:07 vm10 ceph-mon[53712]: pgmap v241: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:07 vm11 ceph-mon[53973]: pgmap v241: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:09.673 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:09.673 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:09 vm10 ceph-mon[53712]: pgmap v242: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:09.700 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:09.701 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:09 vm11 ceph-mon[53973]: pgmap v242: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:11.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:11 vm10 ceph-mon[53712]: pgmap v243: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:11 vm11 ceph-mon[53973]: pgmap v243: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:13 vm10 ceph-mon[53712]: pgmap v244: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:13 vm11 ceph-mon[53973]: pgmap v244: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:14.702 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:14.702 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:14.727 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:14.728 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:15 vm11 ceph-mon[53973]: pgmap v245: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:15 vm10 ceph-mon[53712]: pgmap v245: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:16.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:16 vm11 ceph-mon[53973]: pgmap v246: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:16.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:16 vm10 ceph-mon[53712]: pgmap v246: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:19.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:19 vm10 ceph-mon[53712]: pgmap v247: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:19.729 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:19.730 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:19.756 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:19.757 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:19.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:19 vm11 ceph-mon[53973]: pgmap v247: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:21 vm10 ceph-mon[53712]: pgmap v248: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:21 vm11 ceph-mon[53973]: pgmap v248: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:23.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:23 vm10 ceph-mon[53712]: pgmap v249: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:23 vm11 ceph-mon[53973]: pgmap v249: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:24.758 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:24.759 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:24.784 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:24.784 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:25 vm10 ceph-mon[53712]: pgmap v250: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:25 vm11 ceph-mon[53973]: pgmap v250: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:26.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:26 vm10 ceph-mon[53712]: pgmap v251: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:26 vm11 ceph-mon[53973]: pgmap v251: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:28.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:34:28.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:34:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:34:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:34:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:29 vm10 ceph-mon[53712]: pgmap v252: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:29.786 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:29.786 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:29.813 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:29.814 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:29 vm11 ceph-mon[53973]: pgmap v252: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:30.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:30 vm10 ceph-mon[53712]: pgmap v253: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:30 vm11 ceph-mon[53973]: pgmap v253: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:33 vm10 ceph-mon[53712]: pgmap v254: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:33 vm11 ceph-mon[53973]: pgmap v254: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:34.815 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:34.816 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:34.841 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:34.842 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:35 vm10 ceph-mon[53712]: pgmap v255: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:35 vm11 ceph-mon[53973]: pgmap v255: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:36.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:36 vm10 ceph-mon[53712]: pgmap v256: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:36.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:36 vm11 ceph-mon[53973]: pgmap v256: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:39.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:39 vm10 ceph-mon[53712]: pgmap v257: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:39 vm11 ceph-mon[53973]: pgmap v257: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:39.843 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:39.844 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:39.869 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:39.870 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:40.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:40 vm10 ceph-mon[53712]: pgmap v258: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:40 vm11 ceph-mon[53973]: pgmap v258: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:43 vm10 ceph-mon[53712]: pgmap v259: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:43 vm11 ceph-mon[53973]: pgmap v259: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:44.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:44 vm11 ceph-mon[53973]: pgmap v260: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:44.871 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:44.872 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:44.905 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:44.906 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:44.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:44 vm10 ceph-mon[53712]: pgmap v260: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:47.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:47 vm10 ceph-mon[53712]: pgmap v261: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:47 vm11 ceph-mon[53973]: pgmap v261: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:48 vm11 ceph-mon[53973]: pgmap v262: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:48 vm10 ceph-mon[53712]: pgmap v262: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:49.907 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:49.908 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:50.024 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:50.024 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:51 vm11 ceph-mon[53973]: pgmap v263: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:51 vm10 ceph-mon[53712]: pgmap v263: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:53 vm10 ceph-mon[53712]: pgmap v264: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:34:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:34:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:34:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:34:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:34:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:53 vm11 ceph-mon[53973]: pgmap v264: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:34:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:34:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:34:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:34:53.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:34:54.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:54 vm11 ceph-mon[53973]: pgmap v265: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:54.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:54 vm10 ceph-mon[53712]: pgmap v265: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:55.026 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:34:55.027 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:34:55.062 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:34:55.063 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:34:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:57 vm11 ceph-mon[53973]: pgmap v266: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:57 vm10 ceph-mon[53712]: pgmap v266: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:34:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:34:58 vm11 ceph-mon[53973]: pgmap v267: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:34:58.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:34:58 vm10 ceph-mon[53712]: pgmap v267: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:00.064 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:00.065 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:00.092 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:00.092 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:01 vm10 ceph-mon[53712]: pgmap v268: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:01 vm11 ceph-mon[53973]: pgmap v268: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:03 vm10 ceph-mon[53712]: pgmap v269: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:03 vm11 ceph-mon[53973]: pgmap v269: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:04.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:04 vm11 ceph-mon[53973]: pgmap v270: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:04.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:04 vm10 ceph-mon[53712]: pgmap v270: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:05.094 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:05.094 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:05.120 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:05.121 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:07 vm10 ceph-mon[53712]: pgmap v271: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:07 vm11 ceph-mon[53973]: pgmap v271: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:08 vm11 ceph-mon[53973]: pgmap v272: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:08 vm10 ceph-mon[53712]: pgmap v272: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:10.122 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:10.123 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:10.148 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:10.148 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:11 vm10 ceph-mon[53712]: pgmap v273: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:11 vm11 ceph-mon[53973]: pgmap v273: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:13 vm10 ceph-mon[53712]: pgmap v274: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:13 vm11 ceph-mon[53973]: pgmap v274: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:14.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:14 vm10 ceph-mon[53712]: pgmap v275: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:14.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:14 vm11 ceph-mon[53973]: pgmap v275: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:15.150 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:15.150 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:15.175 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:15.175 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:17 vm11 ceph-mon[53973]: pgmap v276: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:17 vm10 ceph-mon[53712]: pgmap v276: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:18 vm11 ceph-mon[53973]: pgmap v277: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:18 vm10 ceph-mon[53712]: pgmap v277: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:20.176 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:20.177 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:20.202 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:20.202 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:21 vm10 ceph-mon[53712]: pgmap v278: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:21.543 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:21 vm11 ceph-mon[53973]: pgmap v278: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:23.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:23 vm10 ceph-mon[53712]: pgmap v279: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:23 vm11 ceph-mon[53973]: pgmap v279: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:24.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:24 vm10 ceph-mon[53712]: pgmap v280: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:24.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:24 vm11 ceph-mon[53973]: pgmap v280: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:25.204 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:25.204 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:25.243 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:25.243 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:27 vm10 ceph-mon[53712]: pgmap v281: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:27 vm11 ceph-mon[53973]: pgmap v281: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:28.660 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:35:28.660 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:35:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:35:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:35:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:29 vm10 ceph-mon[53712]: pgmap v282: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:29 vm11 ceph-mon[53973]: pgmap v282: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:30.244 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:30.245 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:30.270 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:30.270 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:30.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:30 vm10 ceph-mon[53712]: pgmap v283: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:30 vm11 ceph-mon[53973]: pgmap v283: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:33 vm10 ceph-mon[53712]: pgmap v284: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:33 vm11 ceph-mon[53973]: pgmap v284: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:34.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:34 vm10 ceph-mon[53712]: pgmap v285: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:34.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:34 vm11 ceph-mon[53973]: pgmap v285: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:35.271 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:35.272 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:35.297 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:35.297 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:37 vm10 ceph-mon[53712]: pgmap v286: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:37 vm11 ceph-mon[53973]: pgmap v286: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:38 vm11 ceph-mon[53973]: pgmap v287: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:38 vm10 ceph-mon[53712]: pgmap v287: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:40.298 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:40.299 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:40.325 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:40.326 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:41 vm10 ceph-mon[53712]: pgmap v288: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:41 vm11 ceph-mon[53973]: pgmap v288: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:43 vm10 ceph-mon[53712]: pgmap v289: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:43 vm11 ceph-mon[53973]: pgmap v289: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:45.327 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:45.328 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:45.357 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:45.357 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:45 vm10 ceph-mon[53712]: pgmap v290: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:45 vm11 ceph-mon[53973]: pgmap v290: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:47 vm11 ceph-mon[53973]: pgmap v291: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:47 vm10 ceph-mon[53712]: pgmap v291: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:48.662 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:48 vm10 ceph-mon[53712]: pgmap v292: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:48 vm11 ceph-mon[53973]: pgmap v292: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:50.359 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:50.359 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:50.392 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:50.393 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:51.543 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:51 vm11 ceph-mon[53973]: pgmap v293: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:51.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:51 vm10 ceph-mon[53712]: pgmap v293: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:53 vm11 ceph-mon[53973]: pgmap v294: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:53.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:35:53.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:35:53.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:53 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:35:53.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:53 vm10 ceph-mon[53712]: pgmap v294: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:53.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:35:53.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:35:53.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:53 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:35:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:35:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:35:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:35:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:35:55.394 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:35:55.394 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:35:55.420 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:35:55.420 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:35:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:55 vm11 ceph-mon[53973]: pgmap v295: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:55 vm10 ceph-mon[53712]: pgmap v295: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:57 vm11 ceph-mon[53973]: pgmap v296: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:57 vm10 ceph-mon[53712]: pgmap v296: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:35:58.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:35:58 vm10 ceph-mon[53712]: pgmap v297: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:35:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:35:58 vm11 ceph-mon[53973]: pgmap v297: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:00.422 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:00.422 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:00.448 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:00.448 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:01 vm10 ceph-mon[53712]: pgmap v298: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:01 vm11 ceph-mon[53973]: pgmap v298: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:03.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:03 vm10 ceph-mon[53712]: pgmap v299: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:03 vm11 ceph-mon[53973]: pgmap v299: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:05.450 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:05.450 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:05.476 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:05.476 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:05 vm11 ceph-mon[53973]: pgmap v300: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:05 vm10 ceph-mon[53712]: pgmap v300: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:07 vm11 ceph-mon[53973]: pgmap v301: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:07.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:07 vm10 ceph-mon[53712]: pgmap v301: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:08.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:08 vm10 ceph-mon[53712]: pgmap v302: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:08 vm11 ceph-mon[53973]: pgmap v302: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:10.477 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:10.478 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:10.502 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:10.503 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:11 vm10 ceph-mon[53712]: pgmap v303: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:11 vm11 ceph-mon[53973]: pgmap v303: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:13 vm10 ceph-mon[53712]: pgmap v304: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:13 vm11 ceph-mon[53973]: pgmap v304: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:15 vm10 ceph-mon[53712]: pgmap v305: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:15.504 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:15.505 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:15.530 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:15.530 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:15 vm11 ceph-mon[53973]: pgmap v305: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:17 vm11 ceph-mon[53973]: pgmap v306: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:17 vm10 ceph-mon[53712]: pgmap v306: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:18.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:18 vm10 ceph-mon[53712]: pgmap v307: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:18 vm11 ceph-mon[53973]: pgmap v307: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:20.532 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:20.532 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:20.558 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:20.558 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:21.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:21 vm10 ceph-mon[53712]: pgmap v308: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:21.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:21 vm11 ceph-mon[53973]: pgmap v308: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:23 vm10 ceph-mon[53712]: pgmap v309: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:23 vm11 ceph-mon[53973]: pgmap v309: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:25.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:25 vm10 ceph-mon[53712]: pgmap v310: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:25.559 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:25.560 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:25.586 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:25.586 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:25 vm11 ceph-mon[53973]: pgmap v310: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:27 vm11 ceph-mon[53973]: pgmap v311: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:27 vm10 ceph-mon[53712]: pgmap v311: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:28 vm11 ceph-mon[53973]: pgmap v312: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:36:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:36:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:28 vm10 ceph-mon[53712]: pgmap v312: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:36:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:36:30.588 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:30.588 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:30.612 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:30.613 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:31 vm10 ceph-mon[53712]: pgmap v313: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:31 vm11 ceph-mon[53973]: pgmap v313: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:33 vm10 ceph-mon[53712]: pgmap v314: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:33 vm11 ceph-mon[53973]: pgmap v314: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:35 vm10 ceph-mon[53712]: pgmap v315: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:35 vm11 ceph-mon[53973]: pgmap v315: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:35.614 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:35.614 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:35.640 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:35.641 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:37 vm11 ceph-mon[53973]: pgmap v316: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:37 vm10 ceph-mon[53712]: pgmap v316: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:38 vm11 ceph-mon[53973]: pgmap v317: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:38 vm10 ceph-mon[53712]: pgmap v317: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:40.642 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:40.642 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:40.668 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:40.668 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:41 vm11 ceph-mon[53973]: pgmap v318: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:41 vm10 ceph-mon[53712]: pgmap v318: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:43.550 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:43 vm10 ceph-mon[53712]: pgmap v319: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:43 vm11 ceph-mon[53973]: pgmap v319: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:45 vm11 ceph-mon[53973]: pgmap v320: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:45.669 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:45.670 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:45 vm10 ceph-mon[53712]: pgmap v320: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:45.695 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:45.695 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:47 vm11 ceph-mon[53973]: pgmap v321: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:47 vm10 ceph-mon[53712]: pgmap v321: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:48 vm11 ceph-mon[53973]: pgmap v322: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:48.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:48 vm10 ceph-mon[53712]: pgmap v322: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:50.697 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:50.697 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:50.723 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:50.724 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:51 vm10 ceph-mon[53712]: pgmap v323: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:51 vm11 ceph-mon[53973]: pgmap v323: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:53 vm10 ceph-mon[53712]: pgmap v324: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:53.482 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:53 vm11 ceph-mon[53973]: pgmap v324: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:36:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:36:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:36:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:36:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:36:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:36:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:36:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:36:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:36:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:36:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:36:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:36:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:55 vm11 ceph-mon[53973]: pgmap v325: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:55 vm10 ceph-mon[53712]: pgmap v325: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:55.725 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:36:55.726 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:36:55.751 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:36:55.752 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:36:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:57 vm11 ceph-mon[53973]: pgmap v326: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:57 vm10 ceph-mon[53712]: pgmap v326: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:36:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:36:58 vm11 ceph-mon[53973]: pgmap v327: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:36:58.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:36:58 vm10 ceph-mon[53712]: pgmap v327: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:00.753 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:00.754 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:00.780 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:00.781 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:01 vm10 ceph-mon[53712]: pgmap v328: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:01.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:01 vm11 ceph-mon[53973]: pgmap v328: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:03.553 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:03 vm10 ceph-mon[53712]: pgmap v329: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:03 vm11 ceph-mon[53973]: pgmap v329: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:05 vm11 ceph-mon[53973]: pgmap v330: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:05 vm10 ceph-mon[53712]: pgmap v330: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:05.782 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:05.783 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:05.808 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:05.809 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:07 vm11 ceph-mon[53973]: pgmap v331: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:07 vm10 ceph-mon[53712]: pgmap v331: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:08 vm11 ceph-mon[53973]: pgmap v332: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:08 vm10 ceph-mon[53712]: pgmap v332: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:10.810 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:10.811 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:10.837 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:10.837 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:11 vm10 ceph-mon[53712]: pgmap v333: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:11.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:11 vm11 ceph-mon[53973]: pgmap v333: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:13.554 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:13 vm10 ceph-mon[53712]: pgmap v334: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:13 vm11 ceph-mon[53973]: pgmap v334: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:15 vm11 ceph-mon[53973]: pgmap v335: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:15 vm10 ceph-mon[53712]: pgmap v335: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:15.838 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:15.838 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:15.864 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:15.864 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:17 vm11 ceph-mon[53973]: pgmap v336: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:17 vm10 ceph-mon[53712]: pgmap v336: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:18 vm11 ceph-mon[53973]: pgmap v337: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:18 vm10 ceph-mon[53712]: pgmap v337: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:20.865 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:20.866 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:20.893 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:20.893 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:21 vm10 ceph-mon[53712]: pgmap v338: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:21.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:21 vm11 ceph-mon[53973]: pgmap v338: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:23.554 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:23 vm10 ceph-mon[53712]: pgmap v339: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:23 vm11 ceph-mon[53973]: pgmap v339: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:25 vm11 ceph-mon[53973]: pgmap v340: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:25 vm10 ceph-mon[53712]: pgmap v340: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:25.895 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:25.895 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:25.922 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:25.923 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:27 vm11 ceph-mon[53973]: pgmap v341: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:27 vm10 ceph-mon[53712]: pgmap v341: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:28 vm11 ceph-mon[53973]: pgmap v342: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:37:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:37:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:28 vm10 ceph-mon[53712]: pgmap v342: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:37:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:37:30.924 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:30.924 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:30.950 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:30.951 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:31 vm11 ceph-mon[53973]: pgmap v343: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:31 vm10 ceph-mon[53712]: pgmap v343: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:33.555 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:33 vm10 ceph-mon[53712]: pgmap v344: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:33 vm11 ceph-mon[53973]: pgmap v344: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:35 vm11 ceph-mon[53973]: pgmap v345: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:35 vm10 ceph-mon[53712]: pgmap v345: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:35.952 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:35.953 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:35.981 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:35.982 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:37 vm11 ceph-mon[53973]: pgmap v346: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:37.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:37 vm10 ceph-mon[53712]: pgmap v346: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:38.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:38 vm11 ceph-mon[53973]: pgmap v347: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:38.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:38 vm10 ceph-mon[53712]: pgmap v347: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:40.983 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:40.984 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:41.011 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:41.012 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:41 vm11 ceph-mon[53973]: pgmap v348: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:41.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:41 vm10 ceph-mon[53712]: pgmap v348: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:43.556 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:43 vm10 ceph-mon[53712]: pgmap v349: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:43 vm11 ceph-mon[53973]: pgmap v349: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:45 vm11 ceph-mon[53973]: pgmap v350: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:45 vm10 ceph-mon[53712]: pgmap v350: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:46.013 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:46.014 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:46.039 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:46.040 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:47 vm11 ceph-mon[53973]: pgmap v351: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:47 vm10 ceph-mon[53712]: pgmap v351: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:48 vm11 ceph-mon[53973]: pgmap v352: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:48 vm10 ceph-mon[53712]: pgmap v352: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:51.041 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:51.042 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:51.068 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:51.069 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:51 vm11 ceph-mon[53973]: pgmap v353: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:51 vm10 ceph-mon[53712]: pgmap v353: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:53.558 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:53 vm10 ceph-mon[53712]: pgmap v354: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:53 vm11 ceph-mon[53973]: pgmap v354: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:37:54.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:37:54.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:37:54.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:37:54.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:37:54.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:37:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:37:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:37:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:37:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:37:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:37:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:37:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:55 vm11 ceph-mon[53973]: pgmap v355: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:55 vm10 ceph-mon[53712]: pgmap v355: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:56.070 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:37:56.071 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:37:56.098 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:37:56.098 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:37:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:57 vm11 ceph-mon[53973]: pgmap v356: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:57 vm10 ceph-mon[53712]: pgmap v356: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:37:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:37:58 vm11 ceph-mon[53973]: pgmap v357: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:37:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:37:58 vm10 ceph-mon[53712]: pgmap v357: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:01.100 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:01.100 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:01.126 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:01.126 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:01 vm11 ceph-mon[53973]: pgmap v358: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:01 vm10 ceph-mon[53712]: pgmap v358: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:03.560 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:03 vm10 ceph-mon[53712]: pgmap v359: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:03 vm11 ceph-mon[53973]: pgmap v359: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:05 vm11 ceph-mon[53973]: pgmap v360: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:05 vm10 ceph-mon[53712]: pgmap v360: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:06.128 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:06.128 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:06.154 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:06.155 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:07 vm11 ceph-mon[53973]: pgmap v361: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:07 vm10 ceph-mon[53712]: pgmap v361: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:08 vm11 ceph-mon[53973]: pgmap v362: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:08 vm10 ceph-mon[53712]: pgmap v362: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:11.156 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:11.156 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:11.182 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:11.182 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:11 vm11 ceph-mon[53973]: pgmap v363: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:11 vm10 ceph-mon[53712]: pgmap v363: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:13.561 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:13 vm10 ceph-mon[53712]: pgmap v364: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:13 vm11 ceph-mon[53973]: pgmap v364: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:15 vm11 ceph-mon[53973]: pgmap v365: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:15 vm10 ceph-mon[53712]: pgmap v365: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:16.183 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:16.184 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:16.209 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:16.209 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:17 vm11 ceph-mon[53973]: pgmap v366: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:17 vm10 ceph-mon[53712]: pgmap v366: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:18 vm11 ceph-mon[53973]: pgmap v367: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:18.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:18 vm10 ceph-mon[53712]: pgmap v367: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:21.211 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:21.211 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:21.239 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:21.240 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:21.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:21 vm11 ceph-mon[53973]: pgmap v368: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:21 vm10 ceph-mon[53712]: pgmap v368: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:23.562 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:23 vm10 ceph-mon[53712]: pgmap v369: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:23.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:23 vm11 ceph-mon[53973]: pgmap v369: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:25 vm11 ceph-mon[53973]: pgmap v370: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:25 vm10 ceph-mon[53712]: pgmap v370: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:26.241 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:26.242 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:26.267 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:26.267 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:27 vm11 ceph-mon[53973]: pgmap v371: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:27.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:27 vm10 ceph-mon[53712]: pgmap v371: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:28 vm11 ceph-mon[53973]: pgmap v372: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:38:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:38:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:28 vm10 ceph-mon[53712]: pgmap v372: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:38:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:38:31.269 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:31.270 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:31.297 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:31.297 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:31.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:31 vm11 ceph-mon[53973]: pgmap v373: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:31 vm10 ceph-mon[53712]: pgmap v373: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:33.563 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:33 vm10 ceph-mon[53712]: pgmap v374: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:33 vm11 ceph-mon[53973]: pgmap v374: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:35 vm11 ceph-mon[53973]: pgmap v375: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:35 vm10 ceph-mon[53712]: pgmap v375: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:36.299 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:36.299 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:36.325 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:36.325 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:37 vm11 ceph-mon[53973]: pgmap v376: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:37 vm10 ceph-mon[53712]: pgmap v376: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:38 vm11 ceph-mon[53973]: pgmap v377: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:38 vm10 ceph-mon[53712]: pgmap v377: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:41.327 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:41.327 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:41.353 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:41.354 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:41 vm11 ceph-mon[53973]: pgmap v378: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:41 vm10 ceph-mon[53712]: pgmap v378: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:43.564 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:43 vm10 ceph-mon[53712]: pgmap v379: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:43 vm11 ceph-mon[53973]: pgmap v379: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:45 vm11 ceph-mon[53973]: pgmap v380: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:45 vm10 ceph-mon[53712]: pgmap v380: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:46.355 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:46.356 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:46.382 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:46.382 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:47 vm11 ceph-mon[53973]: pgmap v381: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:47 vm10 ceph-mon[53712]: pgmap v381: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:48 vm11 ceph-mon[53973]: pgmap v382: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:48 vm10 ceph-mon[53712]: pgmap v382: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:51.384 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:51.384 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:51.410 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:51.410 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:51 vm11 ceph-mon[53973]: pgmap v383: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:51.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:51 vm10 ceph-mon[53712]: pgmap v383: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:53.565 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:53 vm10 ceph-mon[53712]: pgmap v384: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:53 vm11 ceph-mon[53973]: pgmap v384: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:38:54.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:38:54.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:38:54.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:54 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:38:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:38:54.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:38:54.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:54 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:38:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:55 vm11 ceph-mon[53973]: pgmap v385: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:38:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:38:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:38:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:55 vm10 ceph-mon[53712]: pgmap v385: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:38:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:38:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:38:56.412 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:38:56.412 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:38:56.439 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:38:56.439 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:38:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:57 vm11 ceph-mon[53973]: pgmap v386: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:57 vm10 ceph-mon[53712]: pgmap v386: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:38:58 vm11 ceph-mon[53973]: pgmap v387: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:38:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:38:58 vm10 ceph-mon[53712]: pgmap v387: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:01.441 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:01.441 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:01.467 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:01.467 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:01 vm11 ceph-mon[53973]: pgmap v388: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:39:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:01 vm10 ceph-mon[53712]: pgmap v388: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:39:03.566 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:03 vm10 ceph-mon[53712]: pgmap v389: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:03 vm11 ceph-mon[53973]: pgmap v389: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:05 vm11 ceph-mon[53973]: pgmap v390: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:39:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:05 vm10 ceph-mon[53712]: pgmap v390: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:39:06.469 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:06.469 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:06.494 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:06.494 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:07.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:07 vm11 ceph-mon[53973]: pgmap v391: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:39:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:07 vm10 ceph-mon[53712]: pgmap v391: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:39:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:08 vm11 ceph-mon[53973]: pgmap v392: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:08 vm10 ceph-mon[53712]: pgmap v392: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:11.496 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:11.496 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:11.521 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:11.522 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:11 vm11 ceph-mon[53973]: pgmap v393: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:11 vm10 ceph-mon[53712]: pgmap v393: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:13.566 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:13 vm10 ceph-mon[53712]: pgmap v394: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:13 vm11 ceph-mon[53973]: pgmap v394: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:15 vm11 ceph-mon[53973]: pgmap v395: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:15 vm10 ceph-mon[53712]: pgmap v395: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:16.523 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:16.523 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:16.550 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:16.550 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:17 vm11 ceph-mon[53973]: pgmap v396: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:17 vm10 ceph-mon[53712]: pgmap v396: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:18 vm11 ceph-mon[53973]: pgmap v397: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:18 vm10 ceph-mon[53712]: pgmap v397: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:21.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:21 vm11 ceph-mon[53973]: pgmap v398: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:21.552 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:21.552 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:21.578 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:21.579 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:21 vm10 ceph-mon[53712]: pgmap v398: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:23.568 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:23 vm10 ceph-mon[53712]: pgmap v399: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:23 vm11 ceph-mon[53973]: pgmap v399: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:25 vm11 ceph-mon[53973]: pgmap v400: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:25.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:25 vm10 ceph-mon[53712]: pgmap v400: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:26.580 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:26.581 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:26.607 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:26.608 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:27 vm11 ceph-mon[53973]: pgmap v401: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:27.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:27 vm10 ceph-mon[53712]: pgmap v401: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:28 vm11 ceph-mon[53973]: pgmap v402: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:39:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:39:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:28 vm10 ceph-mon[53712]: pgmap v402: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:39:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:39:31.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:31 vm11 ceph-mon[53973]: pgmap v403: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:31.609 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:31.610 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:31.635 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:31.636 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:31 vm10 ceph-mon[53712]: pgmap v403: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:33.569 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:33 vm10 ceph-mon[53712]: pgmap v404: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:33 vm11 ceph-mon[53973]: pgmap v404: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:35 vm11 ceph-mon[53973]: pgmap v405: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:35 vm10 ceph-mon[53712]: pgmap v405: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:36.638 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:36.638 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:36.670 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:36.670 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:37 vm11 ceph-mon[53973]: pgmap v406: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:37 vm10 ceph-mon[53712]: pgmap v406: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:38 vm11 ceph-mon[53973]: pgmap v407: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:38 vm10 ceph-mon[53712]: pgmap v407: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:41 vm11 ceph-mon[53973]: pgmap v408: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:41.672 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:41.672 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:41 vm10 ceph-mon[53712]: pgmap v408: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:41.698 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:41.698 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:43.569 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:43 vm10 ceph-mon[53712]: pgmap v409: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:43 vm11 ceph-mon[53973]: pgmap v409: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:45 vm11 ceph-mon[53973]: pgmap v410: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:45 vm10 ceph-mon[53712]: pgmap v410: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:46.700 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:46.700 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:46.728 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:46.729 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:47.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:47 vm11 ceph-mon[53973]: pgmap v411: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:47 vm10 ceph-mon[53712]: pgmap v411: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:48 vm11 ceph-mon[53973]: pgmap v412: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:48 vm10 ceph-mon[53712]: pgmap v412: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:51 vm11 ceph-mon[53973]: pgmap v413: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:51.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:51 vm10 ceph-mon[53712]: pgmap v413: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:51.730 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:51.731 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:51.756 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:51.757 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:53.570 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:53 vm10 ceph-mon[53712]: pgmap v414: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:53 vm11 ceph-mon[53973]: pgmap v414: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:55.259 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:55 vm11 ceph-mon[53973]: pgmap v415: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:55.259 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:39:55.259 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:39:55.259 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:55 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:39:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:55 vm10 ceph-mon[53712]: pgmap v415: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:39:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:39:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:55 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:39:56.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:39:56.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:39:56.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:39:56.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:39:56.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:39:56.759 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:39:56.759 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:39:56.797 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:39:56.798 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:39:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:39:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:39:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:39:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:39:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:39:57.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:57 vm11 ceph-mon[53973]: pgmap v416: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:57 vm10 ceph-mon[53712]: pgmap v416: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:39:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:39:58 vm11 ceph-mon[53973]: pgmap v417: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:39:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:39:58 vm10 ceph-mon[53712]: pgmap v417: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:00 vm11 ceph-mon[53973]: overall HEALTH_OK 2026-03-09T17:40:00.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:00 vm10 ceph-mon[53712]: overall HEALTH_OK 2026-03-09T17:40:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:01 vm11 ceph-mon[53973]: pgmap v418: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:01 vm10 ceph-mon[53712]: pgmap v418: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:01.800 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:01.800 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:01.826 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:01.827 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:03.570 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:03 vm10 ceph-mon[53712]: pgmap v419: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:03 vm11 ceph-mon[53973]: pgmap v419: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:05 vm11 ceph-mon[53973]: pgmap v420: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:05 vm10 ceph-mon[53712]: pgmap v420: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:06.828 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:06.829 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:06.853 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:06.854 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:07 vm11 ceph-mon[53973]: pgmap v421: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:07 vm10 ceph-mon[53712]: pgmap v421: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:08 vm11 ceph-mon[53973]: pgmap v422: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:08.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:08 vm10 ceph-mon[53712]: pgmap v422: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:11 vm11 ceph-mon[53973]: pgmap v423: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:11 vm10 ceph-mon[53712]: pgmap v423: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:11.855 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:11.856 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:11.881 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:11.881 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:13.571 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:13 vm10 ceph-mon[53712]: pgmap v424: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:13 vm11 ceph-mon[53973]: pgmap v424: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:15 vm11 ceph-mon[53973]: pgmap v425: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:15 vm10 ceph-mon[53712]: pgmap v425: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:16.882 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:16.883 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:16.910 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:16.911 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:17 vm11 ceph-mon[53973]: pgmap v426: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:17 vm10 ceph-mon[53712]: pgmap v426: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:18 vm11 ceph-mon[53973]: pgmap v427: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:18 vm10 ceph-mon[53712]: pgmap v427: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:21.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:21 vm11 ceph-mon[53973]: pgmap v428: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:21.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:21 vm10 ceph-mon[53712]: pgmap v428: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:21.913 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:21.914 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:21.942 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:21.942 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:23.573 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:23 vm10 ceph-mon[53712]: pgmap v429: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:23 vm11 ceph-mon[53973]: pgmap v429: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:25 vm11 ceph-mon[53973]: pgmap v430: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:25 vm10 ceph-mon[53712]: pgmap v430: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:26.944 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:26.944 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:26.973 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:26.974 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:27.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:27 vm11 ceph-mon[53973]: pgmap v431: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:27 vm10 ceph-mon[53712]: pgmap v431: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:28 vm11 ceph-mon[53973]: pgmap v432: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:40:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:40:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:28 vm10 ceph-mon[53712]: pgmap v432: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:40:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:40:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:31 vm11 ceph-mon[53973]: pgmap v433: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:31 vm10 ceph-mon[53712]: pgmap v433: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:31.975 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:31.976 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:32.005 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:32.006 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:33.574 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:33 vm10 ceph-mon[53712]: pgmap v434: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:33 vm11 ceph-mon[53973]: pgmap v434: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:35 vm11 ceph-mon[53973]: pgmap v435: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:35 vm10 ceph-mon[53712]: pgmap v435: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:37.008 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:37.008 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:37.034 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:37.034 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:37 vm11 ceph-mon[53973]: pgmap v436: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:37 vm10 ceph-mon[53712]: pgmap v436: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:38 vm11 ceph-mon[53973]: pgmap v437: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:38 vm10 ceph-mon[53712]: pgmap v437: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:41 vm11 ceph-mon[53973]: pgmap v438: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:41 vm10 ceph-mon[53712]: pgmap v438: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:42.036 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:42.036 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:42.061 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:42.062 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:43.575 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:43 vm10 ceph-mon[53712]: pgmap v439: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:43 vm11 ceph-mon[53973]: pgmap v439: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:45 vm11 ceph-mon[53973]: pgmap v440: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:45 vm10 ceph-mon[53712]: pgmap v440: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:47.064 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:47.064 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:47.090 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:47.090 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:47 vm11 ceph-mon[53973]: pgmap v441: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:47 vm10 ceph-mon[53712]: pgmap v441: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:48 vm11 ceph-mon[53973]: pgmap v442: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:48 vm10 ceph-mon[53712]: pgmap v442: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:51 vm11 ceph-mon[53973]: pgmap v443: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:51 vm10 ceph-mon[53712]: pgmap v443: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:52.092 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:52.092 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:52.119 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:52.120 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:53.579 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:53 vm10 ceph-mon[53712]: pgmap v444: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:53 vm11 ceph-mon[53973]: pgmap v444: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:55 vm11 ceph-mon[53973]: pgmap v445: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:55 vm10 ceph-mon[53712]: pgmap v445: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:40:56.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:40:56.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:56 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:40:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:40:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:40:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:56 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:40:57.121 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:40:57.122 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:40:57.209 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:40:57.210 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:40:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:57 vm11 ceph-mon[53973]: pgmap v446: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:40:57.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:40:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:57 vm10 ceph-mon[53712]: pgmap v446: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:40:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:40:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:40:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:40:58 vm11 ceph-mon[53973]: pgmap v447: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:40:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:40:58 vm10 ceph-mon[53712]: pgmap v447: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:01 vm11 ceph-mon[53973]: pgmap v448: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:01 vm10 ceph-mon[53712]: pgmap v448: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:02.211 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:02.211 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:02.241 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:02.242 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:03.577 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:03 vm10 ceph-mon[53712]: pgmap v449: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:03 vm11 ceph-mon[53973]: pgmap v449: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:05 vm10 ceph-mon[53712]: pgmap v450: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:05 vm11 ceph-mon[53973]: pgmap v450: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:07.243 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:07.244 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:07.268 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:07.269 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:07.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:07 vm11 ceph-mon[53973]: pgmap v451: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:07 vm10 ceph-mon[53712]: pgmap v451: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:08 vm11 ceph-mon[53973]: pgmap v452: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:08 vm10 ceph-mon[53712]: pgmap v452: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:11 vm11 ceph-mon[53973]: pgmap v453: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:11 vm10 ceph-mon[53712]: pgmap v453: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:12.270 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:12.271 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:12.299 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:12.300 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:13.578 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:13 vm10 ceph-mon[53712]: pgmap v454: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:13 vm11 ceph-mon[53973]: pgmap v454: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:15 vm11 ceph-mon[53973]: pgmap v455: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:15 vm10 ceph-mon[53712]: pgmap v455: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:17.301 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:17.302 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:17.329 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:17.329 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:17 vm11 ceph-mon[53973]: pgmap v456: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:17 vm10 ceph-mon[53712]: pgmap v456: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:18 vm11 ceph-mon[53973]: pgmap v457: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:18 vm10 ceph-mon[53712]: pgmap v457: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:21.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:21 vm11 ceph-mon[53973]: pgmap v458: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:21.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:21 vm10 ceph-mon[53712]: pgmap v458: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:22.331 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:22.331 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:22.357 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:22.358 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:23.578 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:23 vm10 ceph-mon[53712]: pgmap v459: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:23 vm11 ceph-mon[53973]: pgmap v459: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:25 vm11 ceph-mon[53973]: pgmap v460: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:25 vm10 ceph-mon[53712]: pgmap v460: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:27.359 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:27.360 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:27.385 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:27.386 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:27 vm11 ceph-mon[53973]: pgmap v461: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:27 vm10 ceph-mon[53712]: pgmap v461: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:28 vm11 ceph-mon[53973]: pgmap v462: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:41:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:41:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:28 vm10 ceph-mon[53712]: pgmap v462: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:41:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:41:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:31 vm11 ceph-mon[53973]: pgmap v463: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:31 vm10 ceph-mon[53712]: pgmap v463: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:32.387 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:32.388 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:32.414 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:32.415 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:33.579 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:33 vm10 ceph-mon[53712]: pgmap v464: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:33 vm11 ceph-mon[53973]: pgmap v464: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:35 vm11 ceph-mon[53973]: pgmap v465: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:35 vm10 ceph-mon[53712]: pgmap v465: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:37.416 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:37.417 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:37.442 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:37.443 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:37 vm11 ceph-mon[53973]: pgmap v466: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:37 vm10 ceph-mon[53712]: pgmap v466: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:38 vm11 ceph-mon[53973]: pgmap v467: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:38.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:38 vm10 ceph-mon[53712]: pgmap v467: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:41 vm11 ceph-mon[53973]: pgmap v468: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:41 vm10 ceph-mon[53712]: pgmap v468: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:42.444 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:42.444 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:42.470 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:42.470 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:43.579 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:43 vm10 ceph-mon[53712]: pgmap v469: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:43 vm11 ceph-mon[53973]: pgmap v469: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:45 vm11 ceph-mon[53973]: pgmap v470: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:45 vm10 ceph-mon[53712]: pgmap v470: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:47.472 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:47.472 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:47.497 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:47.497 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:47 vm11 ceph-mon[53973]: pgmap v471: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:47 vm10 ceph-mon[53712]: pgmap v471: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:48 vm11 ceph-mon[53973]: pgmap v472: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:48 vm10 ceph-mon[53712]: pgmap v472: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:51 vm11 ceph-mon[53973]: pgmap v473: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:51 vm10 ceph-mon[53712]: pgmap v473: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:52.499 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:52.499 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:52.525 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:52.525 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:53.580 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:53 vm10 ceph-mon[53712]: pgmap v474: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:53 vm11 ceph-mon[53973]: pgmap v474: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:55 vm11 ceph-mon[53973]: pgmap v475: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:55 vm10 ceph-mon[53712]: pgmap v475: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:57.527 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:41:57.527 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:41:57.554 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:41:57.555 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:41:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:57 vm11 ceph-mon[53973]: pgmap v476: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:41:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:41:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:57 vm10 ceph-mon[53712]: pgmap v476: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:41:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:41:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:41:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:41:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:41:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:41:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:41:58 vm11 ceph-mon[53973]: pgmap v477: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:41:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:41:58 vm10 ceph-mon[53712]: pgmap v477: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:01 vm11 ceph-mon[53973]: pgmap v478: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:01 vm10 ceph-mon[53712]: pgmap v478: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:02.556 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:02.557 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:02.583 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:02.583 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:03.581 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:03 vm10 ceph-mon[53712]: pgmap v479: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:03 vm11 ceph-mon[53973]: pgmap v479: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:05 vm11 ceph-mon[53973]: pgmap v480: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:05 vm10 ceph-mon[53712]: pgmap v480: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:07.584 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:07.585 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:07 vm11 ceph-mon[53973]: pgmap v481: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:07.610 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:07.611 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:07 vm10 ceph-mon[53712]: pgmap v481: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:08 vm11 ceph-mon[53973]: pgmap v482: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:08 vm10 ceph-mon[53712]: pgmap v482: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:11 vm11 ceph-mon[53973]: pgmap v483: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:11 vm10 ceph-mon[53712]: pgmap v483: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:12.613 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:12.613 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:12.640 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:12.641 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:13.582 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:13 vm10 ceph-mon[53712]: pgmap v484: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:13 vm11 ceph-mon[53973]: pgmap v484: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:15 vm11 ceph-mon[53973]: pgmap v485: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:15 vm10 ceph-mon[53712]: pgmap v485: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:17 vm11 ceph-mon[53973]: pgmap v486: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:17.642 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:17.643 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:17.670 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:17.671 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:17 vm10 ceph-mon[53712]: pgmap v486: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:18 vm11 ceph-mon[53973]: pgmap v487: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:18 vm10 ceph-mon[53712]: pgmap v487: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:21 vm11 ceph-mon[53973]: pgmap v488: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:21.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:21 vm10 ceph-mon[53712]: pgmap v488: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:22.672 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:22.673 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:22.699 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:22.699 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:23.583 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:23 vm10 ceph-mon[53712]: pgmap v489: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:23 vm11 ceph-mon[53973]: pgmap v489: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:25 vm11 ceph-mon[53973]: pgmap v490: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:25 vm10 ceph-mon[53712]: pgmap v490: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:27 vm11 ceph-mon[53973]: pgmap v491: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:27 vm10 ceph-mon[53712]: pgmap v491: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:27.700 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:27.701 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:27.727 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:27.727 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:28 vm11 ceph-mon[53973]: pgmap v492: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:42:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:42:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:28 vm10 ceph-mon[53712]: pgmap v492: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:42:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:42:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:31 vm11 ceph-mon[53973]: pgmap v493: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:31 vm10 ceph-mon[53712]: pgmap v493: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:32.729 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:32.730 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:32.757 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:32.758 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:33.584 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:33 vm10 ceph-mon[53712]: pgmap v494: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:33 vm11 ceph-mon[53973]: pgmap v494: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:35 vm11 ceph-mon[53973]: pgmap v495: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:35 vm10 ceph-mon[53712]: pgmap v495: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:37 vm11 ceph-mon[53973]: pgmap v496: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:37.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:37 vm10 ceph-mon[53712]: pgmap v496: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:37.759 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:37.760 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:37.786 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:37.787 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:38.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:38 vm11 ceph-mon[53973]: pgmap v497: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:38.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:38 vm10 ceph-mon[53712]: pgmap v497: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:41 vm11 ceph-mon[53973]: pgmap v498: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:41.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:41 vm10 ceph-mon[53712]: pgmap v498: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:42.788 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:42.789 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:42.814 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:42.815 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:43.585 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:43 vm10 ceph-mon[53712]: pgmap v499: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:43 vm11 ceph-mon[53973]: pgmap v499: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:45 vm11 ceph-mon[53973]: pgmap v500: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:45 vm10 ceph-mon[53712]: pgmap v500: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:47 vm11 ceph-mon[53973]: pgmap v501: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:47 vm10 ceph-mon[53712]: pgmap v501: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:47.816 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:47.817 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:47.842 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:47.842 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:48 vm11 ceph-mon[53973]: pgmap v502: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:48 vm10 ceph-mon[53712]: pgmap v502: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:51 vm11 ceph-mon[53973]: pgmap v503: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:51.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:51 vm10 ceph-mon[53712]: pgmap v503: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:52.844 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:52.844 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:52.870 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:52.871 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:53.586 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:53 vm10 ceph-mon[53712]: pgmap v504: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:53 vm11 ceph-mon[53973]: pgmap v504: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:55.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:55 vm11 ceph-mon[53973]: pgmap v505: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:55 vm10 ceph-mon[53712]: pgmap v505: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:57 vm11 ceph-mon[53973]: pgmap v506: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:42:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:42:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:57 vm10 ceph-mon[53712]: pgmap v506: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:42:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:42:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:42:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:42:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:42:57.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:42:57.873 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:42:57.873 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:42:57.901 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:42:57.902 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:42:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:42:58 vm11 ceph-mon[53973]: pgmap v507: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:42:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:42:58 vm10 ceph-mon[53712]: pgmap v507: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:01 vm11 ceph-mon[53973]: pgmap v508: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:01 vm10 ceph-mon[53712]: pgmap v508: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:02.903 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:02.904 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:02.935 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:02.936 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:03.586 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:03 vm10 ceph-mon[53712]: pgmap v509: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:03 vm11 ceph-mon[53973]: pgmap v509: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:05 vm11 ceph-mon[53973]: pgmap v510: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:05 vm10 ceph-mon[53712]: pgmap v510: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:07 vm11 ceph-mon[53973]: pgmap v511: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:07 vm10 ceph-mon[53712]: pgmap v511: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:07.938 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:07.938 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:07.964 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:07.965 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:08 vm11 ceph-mon[53973]: pgmap v512: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:08 vm10 ceph-mon[53712]: pgmap v512: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:11 vm11 ceph-mon[53973]: pgmap v513: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:11 vm10 ceph-mon[53712]: pgmap v513: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:12.966 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:12.967 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:12.992 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:12.992 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:13.587 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:13 vm10 ceph-mon[53712]: pgmap v514: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:13 vm11 ceph-mon[53973]: pgmap v514: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:15 vm11 ceph-mon[53973]: pgmap v515: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:15 vm10 ceph-mon[53712]: pgmap v515: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:17 vm11 ceph-mon[53973]: pgmap v516: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:17 vm10 ceph-mon[53712]: pgmap v516: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:17.993 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:17.994 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:18.019 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:18.019 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:18 vm11 ceph-mon[53973]: pgmap v517: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:18 vm10 ceph-mon[53712]: pgmap v517: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:21 vm11 ceph-mon[53973]: pgmap v518: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:21 vm10 ceph-mon[53712]: pgmap v518: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:23.021 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:23.021 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:23.046 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:23.047 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:23.588 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:23 vm10 ceph-mon[53712]: pgmap v519: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:23 vm11 ceph-mon[53973]: pgmap v519: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:25 vm11 ceph-mon[53973]: pgmap v520: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:25 vm10 ceph-mon[53712]: pgmap v520: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:27 vm11 ceph-mon[53973]: pgmap v521: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:27 vm10 ceph-mon[53712]: pgmap v521: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:28.048 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:28.049 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:28.121 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:28.122 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:28 vm10 ceph-mon[53712]: pgmap v522: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:43:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:43:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:28 vm11 ceph-mon[53973]: pgmap v522: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:43:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:43:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:31 vm11 ceph-mon[53973]: pgmap v523: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:31 vm10 ceph-mon[53712]: pgmap v523: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:33.123 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:33.124 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:33.150 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:33.150 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:33.589 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:33 vm10 ceph-mon[53712]: pgmap v524: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:33 vm11 ceph-mon[53973]: pgmap v524: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:35 vm11 ceph-mon[53973]: pgmap v525: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:35 vm10 ceph-mon[53712]: pgmap v525: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:37 vm11 ceph-mon[53973]: pgmap v526: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:37.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:37 vm10 ceph-mon[53712]: pgmap v526: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:38.152 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:38.152 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:38.178 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:38.179 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:38 vm11 ceph-mon[53973]: pgmap v527: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:38.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:38 vm10 ceph-mon[53712]: pgmap v527: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:41 vm11 ceph-mon[53973]: pgmap v528: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:41.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:41 vm10 ceph-mon[53712]: pgmap v528: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:43.181 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:43.181 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:43.208 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:43.208 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:43 vm11 ceph-mon[53973]: pgmap v529: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:43.591 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:43 vm10 ceph-mon[53712]: pgmap v529: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:45 vm11 ceph-mon[53973]: pgmap v530: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:45 vm10 ceph-mon[53712]: pgmap v530: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:47 vm11 ceph-mon[53973]: pgmap v531: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:47 vm10 ceph-mon[53712]: pgmap v531: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:48.210 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:48.210 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:48.235 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:48.235 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:48 vm11 ceph-mon[53973]: pgmap v532: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:48 vm10 ceph-mon[53712]: pgmap v532: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:51.544 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:51 vm11 ceph-mon[53973]: pgmap v533: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:51 vm10 ceph-mon[53712]: pgmap v533: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:53.236 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:53.237 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:53.265 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:53.266 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:53 vm11 ceph-mon[53973]: pgmap v534: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:53.591 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:53 vm10 ceph-mon[53712]: pgmap v534: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:55 vm11 ceph-mon[53973]: pgmap v535: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:55 vm10 ceph-mon[53712]: pgmap v535: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:57 vm10 ceph-mon[53712]: pgmap v536: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:57.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:57.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:57.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:57 vm11 ceph-mon[53973]: pgmap v536: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:43:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:43:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:43:57.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:43:58.268 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:43:58.269 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:43:58.324 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:43:58.325 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:43:58.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:43:58.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:43:58.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:43:58.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:43:59.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:43:59 vm11 ceph-mon[53973]: pgmap v537: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:43:59.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:43:59 vm10 ceph-mon[53712]: pgmap v537: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:01 vm11 ceph-mon[53973]: pgmap v538: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:01 vm10 ceph-mon[53712]: pgmap v538: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:03.326 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:03.327 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:03.352 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:03.352 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:03 vm11 ceph-mon[53973]: pgmap v539: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:03 vm10 ceph-mon[53712]: pgmap v539: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:05 vm11 ceph-mon[53973]: pgmap v540: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:05 vm10 ceph-mon[53712]: pgmap v540: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:07 vm11 ceph-mon[53973]: pgmap v541: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:07 vm10 ceph-mon[53712]: pgmap v541: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:08.353 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:08.354 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:08.378 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:08.379 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:08 vm10 ceph-mon[53712]: pgmap v542: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:08 vm11 ceph-mon[53973]: pgmap v542: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:11 vm11 ceph-mon[53973]: pgmap v543: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:11.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:11 vm10 ceph-mon[53712]: pgmap v543: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:13.380 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:13.381 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:13.406 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:13.407 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:13 vm11 ceph-mon[53973]: pgmap v544: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:13.593 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:13 vm10 ceph-mon[53712]: pgmap v544: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:15 vm11 ceph-mon[53973]: pgmap v545: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:15.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:15 vm10 ceph-mon[53712]: pgmap v545: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:17 vm11 ceph-mon[53973]: pgmap v546: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:17 vm10 ceph-mon[53712]: pgmap v546: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:18.408 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:18.409 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:18.434 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:18.435 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:18 vm10 ceph-mon[53712]: pgmap v547: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:18 vm11 ceph-mon[53973]: pgmap v547: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:21.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:21 vm10 ceph-mon[53712]: pgmap v548: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:21.794 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:21 vm11 ceph-mon[53973]: pgmap v548: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:23.436 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:23.436 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:23.462 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:23.462 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:23 vm11 ceph-mon[53973]: pgmap v549: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:23 vm10 ceph-mon[53712]: pgmap v549: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:25 vm11 ceph-mon[53973]: pgmap v550: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:25 vm10 ceph-mon[53712]: pgmap v550: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:26 vm11 ceph-mon[53973]: pgmap v551: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:26 vm10 ceph-mon[53712]: pgmap v551: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:28.464 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:28.464 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:28.490 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:28.491 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:28 vm10 ceph-mon[53712]: pgmap v552: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:44:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:44:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:28 vm11 ceph-mon[53973]: pgmap v552: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:44:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:44:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:31 vm11 ceph-mon[53973]: pgmap v553: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:31 vm10 ceph-mon[53712]: pgmap v553: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:33.492 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:33.492 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:33.517 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:33.518 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:33 vm11 ceph-mon[53973]: pgmap v554: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:33.594 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:33 vm10 ceph-mon[53712]: pgmap v554: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:35 vm11 ceph-mon[53973]: pgmap v555: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:35 vm10 ceph-mon[53712]: pgmap v555: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:37 vm11 ceph-mon[53973]: pgmap v556: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:37 vm10 ceph-mon[53712]: pgmap v556: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:38.519 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:38.520 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:38.544 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:38.545 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:38.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:38 vm10 ceph-mon[53712]: pgmap v557: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:38 vm11 ceph-mon[53973]: pgmap v557: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:41 vm11 ceph-mon[53973]: pgmap v558: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:41 vm10 ceph-mon[53712]: pgmap v558: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:43.546 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:43.547 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:43.572 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:43.573 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:43.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:43 vm11 ceph-mon[53973]: pgmap v559: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:43.596 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:43 vm10 ceph-mon[53712]: pgmap v559: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:45 vm11 ceph-mon[53973]: pgmap v560: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:45 vm10 ceph-mon[53712]: pgmap v560: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:47 vm11 ceph-mon[53973]: pgmap v561: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:47 vm10 ceph-mon[53712]: pgmap v561: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:48.574 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:48.575 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:48.600 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:48.601 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:48 vm10 ceph-mon[53712]: pgmap v562: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:48 vm11 ceph-mon[53973]: pgmap v562: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:51 vm11 ceph-mon[53973]: pgmap v563: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:51 vm10 ceph-mon[53712]: pgmap v563: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:53 vm11 ceph-mon[53973]: pgmap v564: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:53.595 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:53 vm10 ceph-mon[53712]: pgmap v564: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:53.602 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:53.602 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:53.628 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:53.629 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:55 vm11 ceph-mon[53973]: pgmap v565: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:55 vm10 ceph-mon[53712]: pgmap v565: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:56 vm11 ceph-mon[53973]: pgmap v566: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:56 vm10 ceph-mon[53712]: pgmap v566: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:44:57.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:44:57.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:44:57.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:57 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:44:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:44:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:44:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:57 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:44:58.630 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:44:58.631 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:44:58.657 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:44:58.658 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:44:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:44:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:44:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:44:58 vm10 ceph-mon[53712]: pgmap v567: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:44:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:44:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:44:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:44:58 vm11 ceph-mon[53973]: pgmap v567: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:01 vm11 ceph-mon[53973]: pgmap v568: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:01 vm10 ceph-mon[53712]: pgmap v568: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:03 vm11 ceph-mon[53973]: pgmap v569: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:03.596 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:03 vm10 ceph-mon[53712]: pgmap v569: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:03.660 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:03.660 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:03.695 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:03.696 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:05 vm11 ceph-mon[53973]: pgmap v570: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:05 vm10 ceph-mon[53712]: pgmap v570: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:07 vm11 ceph-mon[53973]: pgmap v571: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:07 vm10 ceph-mon[53712]: pgmap v571: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:08.697 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:08.697 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:08.725 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:08.725 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:08 vm11 ceph-mon[53973]: pgmap v572: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:08 vm10 ceph-mon[53712]: pgmap v572: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:11 vm11 ceph-mon[53973]: pgmap v573: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:11.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:11 vm10 ceph-mon[53712]: pgmap v573: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:12 vm10 ceph-mon[53712]: pgmap v574: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:12 vm11 ceph-mon[53973]: pgmap v574: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:13.727 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:13.727 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:13.752 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:13.752 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:15 vm10 ceph-mon[53712]: pgmap v575: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:15 vm11 ceph-mon[53973]: pgmap v575: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:17.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:17 vm11 ceph-mon[53973]: pgmap v576: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:17 vm10 ceph-mon[53712]: pgmap v576: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:18.753 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:18.754 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:18.779 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:18.780 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:18 vm11 ceph-mon[53973]: pgmap v577: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:18 vm10 ceph-mon[53712]: pgmap v577: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:21 vm11 ceph-mon[53973]: pgmap v578: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:21 vm10 ceph-mon[53712]: pgmap v578: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:23 vm11 ceph-mon[53973]: pgmap v579: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:23.598 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:23 vm10 ceph-mon[53712]: pgmap v579: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:23.781 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:23.781 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:23.806 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:23.807 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:25 vm10 ceph-mon[53712]: pgmap v580: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:25 vm11 ceph-mon[53973]: pgmap v580: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:27.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:27 vm11 ceph-mon[53973]: pgmap v581: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:27 vm10 ceph-mon[53712]: pgmap v581: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:28.808 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:28.808 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:28.834 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:28.835 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:28 vm11 ceph-mon[53973]: pgmap v582: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:45:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:45:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:28 vm10 ceph-mon[53712]: pgmap v582: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:45:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:45:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:31 vm11 ceph-mon[53973]: pgmap v583: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:31 vm10 ceph-mon[53712]: pgmap v583: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:33 vm11 ceph-mon[53973]: pgmap v584: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:33.599 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:33 vm10 ceph-mon[53712]: pgmap v584: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:33.836 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:33.836 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:33.862 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:33.862 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:35 vm11 ceph-mon[53973]: pgmap v585: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:35 vm10 ceph-mon[53712]: pgmap v585: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:37 vm11 ceph-mon[53973]: pgmap v586: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:37 vm10 ceph-mon[53712]: pgmap v586: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:38.864 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:38.864 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:38.889 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:38.889 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:38 vm11 ceph-mon[53973]: pgmap v587: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:38 vm10 ceph-mon[53712]: pgmap v587: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:41 vm11 ceph-mon[53973]: pgmap v588: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:41 vm10 ceph-mon[53712]: pgmap v588: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:43 vm11 ceph-mon[53973]: pgmap v589: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:43.600 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:43 vm10 ceph-mon[53712]: pgmap v589: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:43.890 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:43.891 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:43.916 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:43.917 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:45 vm11 ceph-mon[53973]: pgmap v590: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:45 vm10 ceph-mon[53712]: pgmap v590: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:47 vm11 ceph-mon[53973]: pgmap v591: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:47 vm10 ceph-mon[53712]: pgmap v591: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:48.918 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:48.919 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:48.945 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:48.946 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:48 vm11 ceph-mon[53973]: pgmap v592: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:48 vm10 ceph-mon[53712]: pgmap v592: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:51 vm11 ceph-mon[53973]: pgmap v593: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:51 vm10 ceph-mon[53712]: pgmap v593: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:53 vm11 ceph-mon[53973]: pgmap v594: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:53.601 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:53 vm10 ceph-mon[53712]: pgmap v594: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:53.947 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:53.948 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:53.973 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:53.974 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:55 vm11 ceph-mon[53973]: pgmap v595: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:55 vm10 ceph-mon[53712]: pgmap v595: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:57 vm11 ceph-mon[53973]: pgmap v596: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:57.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:57 vm10 ceph-mon[53712]: pgmap v596: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:45:58.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:45:58.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:45:58.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:45:58.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:45:58.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:45:58.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:45:58.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:45:58.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:45:58.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:45:58.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:45:58.975 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:45:58.975 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:45:59.002 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:45:59.003 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:45:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:45:59 vm11 ceph-mon[53973]: pgmap v597: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:45:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:45:59 vm10 ceph-mon[53712]: pgmap v597: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:01 vm11 ceph-mon[53973]: pgmap v598: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:01 vm10 ceph-mon[53712]: pgmap v598: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:03 vm11 ceph-mon[53973]: pgmap v599: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:03.602 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:03 vm10 ceph-mon[53712]: pgmap v599: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:04.004 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:04.004 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:04.029 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:04.030 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:05 vm11 ceph-mon[53973]: pgmap v600: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:05 vm10 ceph-mon[53712]: pgmap v600: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:07 vm11 ceph-mon[53973]: pgmap v601: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:07 vm10 ceph-mon[53712]: pgmap v601: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:09.031 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:09.032 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:09.057 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:09.058 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:08 vm11 ceph-mon[53973]: pgmap v602: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:09.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:08 vm10 ceph-mon[53712]: pgmap v602: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:11 vm11 ceph-mon[53973]: pgmap v603: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:11 vm10 ceph-mon[53712]: pgmap v603: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:13 vm11 ceph-mon[53973]: pgmap v604: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:13.603 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:13 vm10 ceph-mon[53712]: pgmap v604: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:14.059 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:14.060 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:14.086 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:14.086 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:15 vm11 ceph-mon[53973]: pgmap v605: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:15 vm10 ceph-mon[53712]: pgmap v605: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:17 vm11 ceph-mon[53973]: pgmap v606: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:17 vm10 ceph-mon[53712]: pgmap v606: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:19.088 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:19.088 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:18 vm11 ceph-mon[53973]: pgmap v607: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:19.114 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:19.114 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:18 vm10 ceph-mon[53712]: pgmap v607: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:21 vm11 ceph-mon[53973]: pgmap v608: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:21 vm10 ceph-mon[53712]: pgmap v608: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:23 vm11 ceph-mon[53973]: pgmap v609: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:23.604 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:23 vm10 ceph-mon[53712]: pgmap v609: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:24.116 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:24.116 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:24.142 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:24.143 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:25 vm10 ceph-mon[53712]: pgmap v610: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:25 vm11 ceph-mon[53973]: pgmap v610: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:27.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:27 vm11 ceph-mon[53973]: pgmap v611: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:27.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:27 vm10 ceph-mon[53712]: pgmap v611: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:28 vm11 ceph-mon[53973]: pgmap v612: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:46:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:46:29.145 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:29.145 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:29.171 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:29.171 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:28 vm10 ceph-mon[53712]: pgmap v612: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:46:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:46:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:31 vm11 ceph-mon[53973]: pgmap v613: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:31 vm10 ceph-mon[53712]: pgmap v613: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:33 vm11 ceph-mon[53973]: pgmap v614: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:33.605 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:33 vm10 ceph-mon[53712]: pgmap v614: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:34.172 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:34.173 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:34.198 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:34.199 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:35 vm11 ceph-mon[53973]: pgmap v615: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:35 vm10 ceph-mon[53712]: pgmap v615: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:37.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:37 vm11 ceph-mon[53973]: pgmap v616: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:37 vm10 ceph-mon[53712]: pgmap v616: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:38 vm11 ceph-mon[53973]: pgmap v617: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:38 vm10 ceph-mon[53712]: pgmap v617: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:39.200 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:39.201 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:39.227 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:39.228 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:41 vm11 ceph-mon[53973]: pgmap v618: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:41 vm10 ceph-mon[53712]: pgmap v618: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:43 vm11 ceph-mon[53973]: pgmap v619: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:43.606 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:43 vm10 ceph-mon[53712]: pgmap v619: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:44.229 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:44.229 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:44.255 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:44.255 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:45 vm11 ceph-mon[53973]: pgmap v620: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:45 vm10 ceph-mon[53712]: pgmap v620: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:47 vm11 ceph-mon[53973]: pgmap v621: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:47 vm10 ceph-mon[53712]: pgmap v621: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:48 vm11 ceph-mon[53973]: pgmap v622: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:48 vm10 ceph-mon[53712]: pgmap v622: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:49.257 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:49.257 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:49.282 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:49.283 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:51 vm11 ceph-mon[53973]: pgmap v623: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:51 vm10 ceph-mon[53712]: pgmap v623: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:53 vm11 ceph-mon[53973]: pgmap v624: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:53.606 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:53 vm10 ceph-mon[53712]: pgmap v624: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:54.284 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:54.285 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:54.310 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:54.311 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:46:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:55 vm11 ceph-mon[53973]: pgmap v625: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:55 vm10 ceph-mon[53712]: pgmap v625: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:57 vm11 ceph-mon[53973]: pgmap v626: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:57 vm10 ceph-mon[53712]: pgmap v626: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:46:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:46:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:46:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:46:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:46:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:46:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:46:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:46:58 vm11 ceph-mon[53973]: pgmap v627: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:46:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:46:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:46:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:46:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:46:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:46:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:46:58 vm10 ceph-mon[53712]: pgmap v627: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:46:59.312 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:46:59.313 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:46:59.337 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:46:59.338 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:01 vm11 ceph-mon[53973]: pgmap v628: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:01 vm10 ceph-mon[53712]: pgmap v628: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:03 vm11 ceph-mon[53973]: pgmap v629: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:03.606 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:03 vm10 ceph-mon[53712]: pgmap v629: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:04.339 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:04.340 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:04.366 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:04.366 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:05 vm11 ceph-mon[53973]: pgmap v630: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:05 vm10 ceph-mon[53712]: pgmap v630: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:07 vm11 ceph-mon[53973]: pgmap v631: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:07 vm10 ceph-mon[53712]: pgmap v631: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:08 vm11 ceph-mon[53973]: pgmap v632: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:09.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:08 vm10 ceph-mon[53712]: pgmap v632: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:09.368 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:09.369 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:09.395 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:09.395 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:11 vm11 ceph-mon[53973]: pgmap v633: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:11 vm10 ceph-mon[53712]: pgmap v633: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:13 vm11 ceph-mon[53973]: pgmap v634: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:13.607 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:13 vm10 ceph-mon[53712]: pgmap v634: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:14.397 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:14.398 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:14.423 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:14.424 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:15 vm11 ceph-mon[53973]: pgmap v635: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:15 vm10 ceph-mon[53712]: pgmap v635: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:17 vm11 ceph-mon[53973]: pgmap v636: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:17 vm10 ceph-mon[53712]: pgmap v636: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:18 vm11 ceph-mon[53973]: pgmap v637: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:18 vm10 ceph-mon[53712]: pgmap v637: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:19.425 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:19.426 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:19.453 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:19.453 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:21 vm11 ceph-mon[53973]: pgmap v638: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:21.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:21 vm10 ceph-mon[53712]: pgmap v638: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:23 vm11 ceph-mon[53973]: pgmap v639: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:23.608 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:23 vm10 ceph-mon[53712]: pgmap v639: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:24.455 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:24.456 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:24.484 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:24.485 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:25 vm11 ceph-mon[53973]: pgmap v640: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:25 vm10 ceph-mon[53712]: pgmap v640: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:27 vm11 ceph-mon[53973]: pgmap v641: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:27 vm10 ceph-mon[53712]: pgmap v641: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:28 vm11 ceph-mon[53973]: pgmap v642: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:47:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:47:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:28 vm10 ceph-mon[53712]: pgmap v642: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:47:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:47:29.486 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:29.487 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:29.512 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:29.513 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:31 vm11 ceph-mon[53973]: pgmap v643: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:31 vm10 ceph-mon[53712]: pgmap v643: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:33 vm11 ceph-mon[53973]: pgmap v644: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:33.609 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:33 vm10 ceph-mon[53712]: pgmap v644: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:34.514 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:34.515 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:34.540 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:34.541 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:35 vm11 ceph-mon[53973]: pgmap v645: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:35 vm10 ceph-mon[53712]: pgmap v645: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:37.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:37 vm11 ceph-mon[53973]: pgmap v646: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:37.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:37 vm10 ceph-mon[53712]: pgmap v646: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:38 vm11 ceph-mon[53973]: pgmap v647: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:38 vm10 ceph-mon[53712]: pgmap v647: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:39.542 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:39.542 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:39.568 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:39.568 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:41 vm11 ceph-mon[53973]: pgmap v648: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:41 vm10 ceph-mon[53712]: pgmap v648: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:43 vm11 ceph-mon[53973]: pgmap v649: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:43.610 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:43 vm10 ceph-mon[53712]: pgmap v649: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:44.570 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:44.570 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:44.595 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:44.596 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:45 vm11 ceph-mon[53973]: pgmap v650: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:45 vm10 ceph-mon[53712]: pgmap v650: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:47 vm11 ceph-mon[53973]: pgmap v651: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:47 vm10 ceph-mon[53712]: pgmap v651: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:48 vm11 ceph-mon[53973]: pgmap v652: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:48 vm10 ceph-mon[53712]: pgmap v652: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:49.598 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:49.598 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:49.625 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:49.626 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:51 vm11 ceph-mon[53973]: pgmap v653: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:51 vm10 ceph-mon[53712]: pgmap v653: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:53 vm11 ceph-mon[53973]: pgmap v654: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:53.610 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:53 vm10 ceph-mon[53712]: pgmap v654: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:54.627 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:54.628 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:54.653 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:54.654 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:47:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:55 vm11 ceph-mon[53973]: pgmap v655: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:55 vm10 ceph-mon[53712]: pgmap v655: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:57 vm11 ceph-mon[53973]: pgmap v656: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:57 vm10 ceph-mon[53712]: pgmap v656: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:47:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:47:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:47:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:47:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:58 vm11 ceph-mon[53973]: pgmap v657: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:47:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:47:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:47:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:47:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:47:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:47:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:47:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:58 vm10 ceph-mon[53712]: pgmap v657: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:47:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:47:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:47:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:47:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:47:59.655 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:47:59.656 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:47:59.682 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:47:59.683 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:01 vm11 ceph-mon[53973]: pgmap v658: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:01 vm10 ceph-mon[53712]: pgmap v658: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:03 vm11 ceph-mon[53973]: pgmap v659: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:03.612 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:03 vm10 ceph-mon[53712]: pgmap v659: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:04.684 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:04.685 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:04.712 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:04.713 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:05 vm11 ceph-mon[53973]: pgmap v660: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:05 vm10 ceph-mon[53712]: pgmap v660: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:07 vm11 ceph-mon[53973]: pgmap v661: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:07 vm10 ceph-mon[53712]: pgmap v661: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:08 vm11 ceph-mon[53973]: pgmap v662: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:08 vm10 ceph-mon[53712]: pgmap v662: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:09.714 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:09.715 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:09.741 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:09.742 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:11 vm11 ceph-mon[53973]: pgmap v663: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:11 vm10 ceph-mon[53712]: pgmap v663: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:13 vm11 ceph-mon[53973]: pgmap v664: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:13.612 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:13 vm10 ceph-mon[53712]: pgmap v664: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:14.743 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:14.744 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:14.770 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:14.771 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:15 vm11 ceph-mon[53973]: pgmap v665: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:15 vm10 ceph-mon[53712]: pgmap v665: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:17.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:17 vm11 ceph-mon[53973]: pgmap v666: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:17 vm10 ceph-mon[53712]: pgmap v666: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:18 vm11 ceph-mon[53973]: pgmap v667: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:18 vm10 ceph-mon[53712]: pgmap v667: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:19.772 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:19.772 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:19.798 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:19.798 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:21 vm11 ceph-mon[53973]: pgmap v668: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:21 vm10 ceph-mon[53712]: pgmap v668: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:23 vm11 ceph-mon[53973]: pgmap v669: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:23.613 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:23 vm10 ceph-mon[53712]: pgmap v669: 97 pgs: 97 active+clean; 453 KiB data, 53 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:24.799 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:24.800 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:24.825 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:24.825 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:25 vm11 ceph-mon[53973]: pgmap v670: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:25 vm10 ceph-mon[53712]: pgmap v670: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:27 vm11 ceph-mon[53973]: pgmap v671: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:27 vm10 ceph-mon[53712]: pgmap v671: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:28 vm11 ceph-mon[53973]: pgmap v672: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:48:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:48:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:28 vm10 ceph-mon[53712]: pgmap v672: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:48:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:48:29.826 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:29.827 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:29.852 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:29.852 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:31 vm11 ceph-mon[53973]: pgmap v673: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:31 vm10 ceph-mon[53712]: pgmap v673: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:33 vm11 ceph-mon[53973]: pgmap v674: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:33.613 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:33 vm10 ceph-mon[53712]: pgmap v674: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:34.853 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:34.854 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:34.880 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:34.880 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:35 vm11 ceph-mon[53973]: pgmap v675: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:35 vm10 ceph-mon[53712]: pgmap v675: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:37 vm11 ceph-mon[53973]: pgmap v676: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:37 vm10 ceph-mon[53712]: pgmap v676: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:38 vm11 ceph-mon[53973]: pgmap v677: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:38 vm10 ceph-mon[53712]: pgmap v677: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:39.882 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:39.882 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:39.909 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:39.909 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:41 vm11 ceph-mon[53973]: pgmap v678: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:41 vm10 ceph-mon[53712]: pgmap v678: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:43 vm11 ceph-mon[53973]: pgmap v679: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:43.614 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:43 vm10 ceph-mon[53712]: pgmap v679: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:44.911 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:44.911 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:44.937 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:44.938 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:45 vm11 ceph-mon[53973]: pgmap v680: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:45 vm10 ceph-mon[53712]: pgmap v680: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:47 vm11 ceph-mon[53973]: pgmap v681: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:47 vm10 ceph-mon[53712]: pgmap v681: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:48 vm11 ceph-mon[53973]: pgmap v682: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:48 vm10 ceph-mon[53712]: pgmap v682: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:49.939 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:49.940 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:49.965 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:49.966 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:51 vm11 ceph-mon[53973]: pgmap v683: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:51 vm10 ceph-mon[53712]: pgmap v683: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:53 vm11 ceph-mon[53973]: pgmap v684: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:53.615 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:53 vm10 ceph-mon[53712]: pgmap v684: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:54.967 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:54.968 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:48:54.994 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:48:54.995 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:48:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:55 vm11 ceph-mon[53973]: pgmap v685: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:55 vm10 ceph-mon[53712]: pgmap v685: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:57 vm11 ceph-mon[53973]: pgmap v686: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:57 vm10 ceph-mon[53712]: pgmap v686: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:48:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:58 vm11 ceph-mon[53973]: pgmap v687: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:48:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:48:59.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:58 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:48:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:58 vm10 ceph-mon[53712]: pgmap v687: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:48:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:48:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:48:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:58 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:48:59.996 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:48:59.997 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:00.024 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:00.025 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:00.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:59 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:49:00.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:59 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:49:00.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:48:59 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:49:00.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:59 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:49:00.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:59 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:49:00.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:48:59 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:49:01.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:00 vm10 ceph-mon[53712]: pgmap v688: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:00 vm11 ceph-mon[53973]: pgmap v688: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:03 vm11 ceph-mon[53973]: pgmap v689: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:03.615 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:03 vm10 ceph-mon[53712]: pgmap v689: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:05.026 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:05.026 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:05.053 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:05.054 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:05 vm11 ceph-mon[53973]: pgmap v690: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:05 vm10 ceph-mon[53712]: pgmap v690: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:07.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:07 vm11 ceph-mon[53973]: pgmap v691: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:07 vm10 ceph-mon[53712]: pgmap v691: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:08 vm11 ceph-mon[53973]: pgmap v692: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:08 vm10 ceph-mon[53712]: pgmap v692: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:10.055 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:10.055 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:10.081 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:10.081 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:11 vm11 ceph-mon[53973]: pgmap v693: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:11 vm10 ceph-mon[53712]: pgmap v693: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:13 vm11 ceph-mon[53973]: pgmap v694: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:13.616 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:13 vm10 ceph-mon[53712]: pgmap v694: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:15.082 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:15.083 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:15.111 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:15.111 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:15 vm10 ceph-mon[53712]: pgmap v695: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:15 vm11 ceph-mon[53973]: pgmap v695: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:17.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:17 vm11 ceph-mon[53973]: pgmap v696: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:17 vm10 ceph-mon[53712]: pgmap v696: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:18 vm11 ceph-mon[53973]: pgmap v697: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:18 vm10 ceph-mon[53712]: pgmap v697: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:20.113 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:20.113 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:20.138 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:20.139 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:21 vm11 ceph-mon[53973]: pgmap v698: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:21 vm10 ceph-mon[53712]: pgmap v698: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:23 vm11 ceph-mon[53973]: pgmap v699: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:23.617 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:23 vm10 ceph-mon[53712]: pgmap v699: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:25.140 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:25.141 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:25.166 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:25.167 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:25 vm10 ceph-mon[53712]: pgmap v700: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:25 vm11 ceph-mon[53973]: pgmap v700: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:27.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:27 vm11 ceph-mon[53973]: pgmap v701: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:27 vm10 ceph-mon[53712]: pgmap v701: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:28 vm11 ceph-mon[53973]: pgmap v702: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:49:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:49:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:28 vm10 ceph-mon[53712]: pgmap v702: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:49:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:49:30.168 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:30.169 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:30.194 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:30.195 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:31 vm11 ceph-mon[53973]: pgmap v703: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:31 vm10 ceph-mon[53712]: pgmap v703: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:33.618 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:33 vm10 ceph-mon[53712]: pgmap v704: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:33 vm11 ceph-mon[53973]: pgmap v704: 97 pgs: 97 active+clean; 453 KiB data, 57 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:35.196 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:35.197 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:35.223 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:35.224 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:35 vm10 ceph-mon[53712]: pgmap v705: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:35 vm11 ceph-mon[53973]: pgmap v705: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:37.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:37 vm11 ceph-mon[53973]: pgmap v706: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:37 vm10 ceph-mon[53712]: pgmap v706: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:38 vm11 ceph-mon[53973]: pgmap v707: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:38 vm10 ceph-mon[53712]: pgmap v707: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:40.226 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:40.226 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:40.251 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:40.252 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:41 vm11 ceph-mon[53973]: pgmap v708: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:41 vm10 ceph-mon[53712]: pgmap v708: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:43 vm11 ceph-mon[53973]: pgmap v709: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:43.618 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:43 vm10 ceph-mon[53712]: pgmap v709: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:45.253 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:45.253 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:45.280 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:45.280 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:45 vm10 ceph-mon[53712]: pgmap v710: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:45 vm11 ceph-mon[53973]: pgmap v710: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:47.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:47 vm11 ceph-mon[53973]: pgmap v711: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:47 vm10 ceph-mon[53712]: pgmap v711: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:49.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:48 vm10 ceph-mon[53712]: pgmap v712: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:48 vm11 ceph-mon[53973]: pgmap v712: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:50.282 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:50.282 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:50.311 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:50.311 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:51.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:51 vm11 ceph-mon[53973]: pgmap v713: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:51 vm10 ceph-mon[53712]: pgmap v713: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:53 vm10 ceph-mon[53712]: pgmap v714: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:53 vm11 ceph-mon[53973]: pgmap v714: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:55.312 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:49:55.313 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:49:55.338 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:49:55.339 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:49:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:55 vm10 ceph-mon[53712]: pgmap v715: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:55 vm11 ceph-mon[53973]: pgmap v715: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:57.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:57 vm11 ceph-mon[53973]: pgmap v716: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:57 vm10 ceph-mon[53712]: pgmap v716: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:49:59.011 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:58 vm10 ceph-mon[53712]: pgmap v717: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:58 vm11 ceph-mon[53973]: pgmap v717: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:49:59.777 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:59 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:49:59.777 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:59 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:49:59.777 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:49:59 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:49:59.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:59 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:49:59.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:59 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:49:59.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:49:59 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:50:00.345 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:00.346 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:50:00.399 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:00.400 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:50:01.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:01 vm11 ceph-mon[53973]: overall HEALTH_OK 2026-03-09T17:50:01.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:50:01.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:01 vm11 ceph-mon[53973]: pgmap v718: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:01.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:50:01.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:50:01.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:50:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:50:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:01 vm10 ceph-mon[53712]: overall HEALTH_OK 2026-03-09T17:50:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:50:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:01 vm10 ceph-mon[53712]: pgmap v718: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:50:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:50:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:01 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:50:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:03 vm11 ceph-mon[53973]: pgmap v719: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:03.620 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:03 vm10 ceph-mon[53712]: pgmap v719: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:05.401 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:05.402 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:50:05.427 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:05.427 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:05 vm10 ceph-mon[53712]: pgmap v720: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:05 vm11 ceph-mon[53973]: pgmap v720: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:07.664 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:07 vm11 ceph-mon[53973]: pgmap v721: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:07 vm10 ceph-mon[53712]: pgmap v721: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:08 vm11 ceph-mon[53973]: pgmap v722: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:08 vm10 ceph-mon[53712]: pgmap v722: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:10.429 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:10.429 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:50:10.458 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:10.458 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:11 vm11 ceph-mon[53973]: pgmap v723: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:11 vm10 ceph-mon[53712]: pgmap v723: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:13 vm10 ceph-mon[53712]: pgmap v724: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:13 vm11 ceph-mon[53973]: pgmap v724: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:15.460 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:15.460 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:50:15.497 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:15.498 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:15 vm10 ceph-mon[53712]: pgmap v725: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:15 vm11 ceph-mon[53973]: pgmap v725: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:17 vm11 ceph-mon[53973]: pgmap v726: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:17 vm10 ceph-mon[53712]: pgmap v726: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:19.292 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:19 vm10 ceph-mon[53712]: pgmap v727: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:19 vm11 ceph-mon[53973]: pgmap v727: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:20.500 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:20.500 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:50:20.560 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:20.561 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:20.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:20 vm10 ceph-mon[53712]: pgmap v728: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:21.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:20 vm11 ceph-mon[53973]: pgmap v728: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:23.622 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:23 vm10 ceph-mon[53712]: pgmap v729: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:23.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:23 vm11 ceph-mon[53973]: pgmap v729: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:25.563 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:25.563 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:50:25.591 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:25.591 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:25 vm10 ceph-mon[53712]: pgmap v730: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:25 vm11 ceph-mon[53973]: pgmap v730: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:27.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:27 vm11 ceph-mon[53973]: pgmap v731: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:27 vm10 ceph-mon[53712]: pgmap v731: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:50:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:50:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:28 vm11 ceph-mon[53973]: pgmap v732: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:50:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:28 vm10 ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:50:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:28 vm10 ceph-mon[53712]: pgmap v732: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:30.593 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:30.593 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10:/fake /mnt/foo -o sync 2026-03-09T17:50:30.618 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:30.619 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:31 vm10 ceph-mon[53712]: pgmap v733: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:31.792 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:31 vm11 ceph-mon[53973]: pgmap v733: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:33.623 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:33 vm10 ceph-mon[53712]: pgmap v734: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:33 vm11 ceph-mon[53973]: pgmap v734: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:35.620 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:35.621 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:50:35.650 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:35.651 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:35 vm10.local ceph-mon[53712]: pgmap v735: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:35 vm11 ceph-mon[53973]: pgmap v735: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:37.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:37 vm11 ceph-mon[53973]: pgmap v736: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:37 vm10.local ceph-mon[53712]: pgmap v736: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:38 vm11 ceph-mon[53973]: pgmap v737: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:38 vm10.local ceph-mon[53712]: pgmap v737: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:40.653 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:40.653 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:50:40.678 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:40.679 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:41 vm10.local ceph-mon[53712]: pgmap v738: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:41 vm11 ceph-mon[53973]: pgmap v738: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:43.623 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:43 vm10.local ceph-mon[53712]: pgmap v739: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:43 vm11 ceph-mon[53973]: pgmap v739: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:45.680 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:45.681 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:50:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:45 vm10.local ceph-mon[53712]: pgmap v740: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:45.730 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:45.730 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:45 vm11 ceph-mon[53973]: pgmap v740: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:47.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:47 vm11 ceph-mon[53973]: pgmap v741: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:47 vm10.local ceph-mon[53712]: pgmap v741: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:48 vm11 ceph-mon[53973]: pgmap v742: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:48 vm10.local ceph-mon[53712]: pgmap v742: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:50.733 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:50.733 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:50:50.909 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:50.910 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:51 vm10.local ceph-mon[53712]: pgmap v743: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:51.793 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:51 vm11 ceph-mon[53973]: pgmap v743: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:53.624 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:53 vm10.local ceph-mon[53712]: pgmap v744: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:53 vm11 ceph-mon[53973]: pgmap v744: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:55 vm10.local ceph-mon[53712]: pgmap v745: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:55 vm11 ceph-mon[53973]: pgmap v745: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:55.912 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:50:55.913 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:50:55.942 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:50:55.943 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:50:57.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:57 vm11 ceph-mon[53973]: pgmap v746: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:57 vm10.local ceph-mon[53712]: pgmap v746: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:50:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:50:58 vm11 ceph-mon[53973]: pgmap v747: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:50:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:50:58 vm10.local ceph-mon[53712]: pgmap v747: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:00.944 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:00.945 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:00.980 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:00.980 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:01 vm10.local ceph-mon[53712]: pgmap v748: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:51:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:51:01.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:51:01.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:51:01.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:51:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:01 vm11 ceph-mon[53973]: pgmap v748: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:51:01.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:51:01.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:51:01.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:51:01.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:01 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:51:03.626 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:03 vm10.local ceph-mon[53712]: pgmap v749: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:03 vm11 ceph-mon[53973]: pgmap v749: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:05.366 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:05 vm10.local ceph-mon[53712]: pgmap v750: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:05 vm11 ceph-mon[53973]: pgmap v750: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:05.982 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:05.982 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:06.010 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:06.010 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:07.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:07 vm11 ceph-mon[53973]: pgmap v751: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:07 vm10.local ceph-mon[53712]: pgmap v751: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:08 vm11 ceph-mon[53973]: pgmap v752: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:08 vm10.local ceph-mon[53712]: pgmap v752: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:11.012 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:11.012 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:11.038 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:11.038 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:11 vm10.local ceph-mon[53712]: pgmap v753: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:11 vm11 ceph-mon[53973]: pgmap v753: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:13 vm10.local ceph-mon[53712]: pgmap v754: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:13 vm11 ceph-mon[53973]: pgmap v754: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:15 vm10.local ceph-mon[53712]: pgmap v755: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:15 vm11 ceph-mon[53973]: pgmap v755: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:16.039 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:16.040 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:16.066 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:16.066 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:17 vm11 ceph-mon[53973]: pgmap v756: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:17 vm10.local ceph-mon[53712]: pgmap v756: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:18 vm11 ceph-mon[53973]: pgmap v757: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:18 vm10.local ceph-mon[53712]: pgmap v757: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:21.068 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:21.069 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:21.094 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:21.094 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:21 vm10.local ceph-mon[53712]: pgmap v758: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:21.793 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:21 vm11 ceph-mon[53973]: pgmap v758: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:23.628 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:23 vm10.local ceph-mon[53712]: pgmap v759: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:23 vm11 ceph-mon[53973]: pgmap v759: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:25.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:25 vm10.local ceph-mon[53712]: pgmap v760: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:25 vm11 ceph-mon[53973]: pgmap v760: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:26.096 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:26.096 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:26.122 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:26.122 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:27.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:27 vm11 ceph-mon[53973]: pgmap v761: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:27 vm10.local ceph-mon[53712]: pgmap v761: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:51:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:51:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:28 vm10.local ceph-mon[53712]: pgmap v762: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:51:29.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:28 vm11 ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:51:29.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:28 vm11 ceph-mon[53973]: pgmap v762: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:31.124 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:31.124 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:31.150 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:31.150 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:31.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:31 vm10.local ceph-mon[53712]: pgmap v763: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:31 vm11 ceph-mon[53973]: pgmap v763: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:33.629 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:33 vm10.local ceph-mon[53712]: pgmap v764: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:33 vm11 ceph-mon[53973]: pgmap v764: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:35 vm10.local ceph-mon[53712]: pgmap v765: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:35 vm11 ceph-mon[53973]: pgmap v765: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:36.152 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:36.152 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:36.179 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:36.179 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:37.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:37 vm11 ceph-mon[53973]: pgmap v766: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:37 vm10.local ceph-mon[53712]: pgmap v766: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:38 vm10.local ceph-mon[53712]: pgmap v767: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:38 vm11 ceph-mon[53973]: pgmap v767: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:41.181 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:41.181 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:41.207 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:41.208 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:41 vm10.local ceph-mon[53712]: pgmap v768: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:41 vm11 ceph-mon[53973]: pgmap v768: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:43.631 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:43 vm10.local ceph-mon[53712]: pgmap v769: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:43 vm11 ceph-mon[53973]: pgmap v769: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:45 vm10.local ceph-mon[53712]: pgmap v770: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:45 vm11 ceph-mon[53973]: pgmap v770: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:46.209 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:46.209 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:46.273 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:46.274 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:47.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:47 vm11 ceph-mon[53973]: pgmap v771: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:47 vm10.local ceph-mon[53712]: pgmap v771: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:48 vm10.local ceph-mon[53712]: pgmap v772: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:48 vm11 ceph-mon[53973]: pgmap v772: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:51.275 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:51.275 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:51.301 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:51.301 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:51 vm10.local ceph-mon[53712]: pgmap v773: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:51 vm11 ceph-mon[53973]: pgmap v773: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:53.631 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:53 vm10.local ceph-mon[53712]: pgmap v774: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:53 vm11 ceph-mon[53973]: pgmap v774: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:55 vm10.local ceph-mon[53712]: pgmap v775: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:55 vm11 ceph-mon[53973]: pgmap v775: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:56.302 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:51:56.303 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:51:56.329 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:51:56.330 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:51:57.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:57 vm11 ceph-mon[53973]: pgmap v776: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:57 vm10.local ceph-mon[53712]: pgmap v776: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:51:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:51:58 vm10.local ceph-mon[53712]: pgmap v777: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:51:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:51:58 vm11 ceph-mon[53973]: pgmap v777: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:01.332 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:01.332 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:01.358 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:01.359 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:01 vm10.local ceph-mon[53712]: pgmap v778: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:52:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:52:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:52:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:52:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:52:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:01 vm11.local ceph-mon[53973]: pgmap v778: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:01 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:52:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:01 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:52:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:01 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:52:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:01 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:52:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:01 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:52:03.633 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:03 vm10.local ceph-mon[53712]: pgmap v779: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:03 vm11.local ceph-mon[53973]: pgmap v779: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:05 vm10.local ceph-mon[53712]: pgmap v780: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:05 vm11.local ceph-mon[53973]: pgmap v780: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:06.361 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:06.361 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:06.388 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:06.388 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:07.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:07 vm11.local ceph-mon[53973]: pgmap v781: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:07 vm10.local ceph-mon[53712]: pgmap v781: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:08 vm10.local ceph-mon[53712]: pgmap v782: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:08 vm11.local ceph-mon[53973]: pgmap v782: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:11.389 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:11.390 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:11.417 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:11.418 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:11 vm10.local ceph-mon[53712]: pgmap v783: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:11 vm11.local ceph-mon[53973]: pgmap v783: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:13.634 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:13 vm10.local ceph-mon[53712]: pgmap v784: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:13 vm11.local ceph-mon[53973]: pgmap v784: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:15.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:15 vm10.local ceph-mon[53712]: pgmap v785: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:52:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:15 vm11.local ceph-mon[53973]: pgmap v785: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:52:16.419 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:16.420 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:16.445 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:16.446 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:17.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:17 vm11.local ceph-mon[53973]: pgmap v786: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:52:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:17 vm10.local ceph-mon[53712]: pgmap v786: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:52:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:18 vm10.local ceph-mon[53712]: pgmap v787: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:18 vm11.local ceph-mon[53973]: pgmap v787: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:21.448 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:21.448 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:21.475 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:21.476 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:21 vm10.local ceph-mon[53712]: pgmap v788: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:52:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:21 vm11.local ceph-mon[53973]: pgmap v788: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T17:52:23.634 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:23 vm10.local ceph-mon[53712]: pgmap v789: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:23 vm11.local ceph-mon[53973]: pgmap v789: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:25 vm11.local ceph-mon[53973]: pgmap v790: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:25 vm10.local ceph-mon[53712]: pgmap v790: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:26.477 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:26.478 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:26.504 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:26.504 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:27 vm11.local ceph-mon[53973]: pgmap v791: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:27 vm10.local ceph-mon[53712]: pgmap v791: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:52:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:52:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:28 vm10.local ceph-mon[53712]: pgmap v792: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:52:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:52:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:28 vm11.local ceph-mon[53973]: pgmap v792: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:31.506 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:31.506 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:31.532 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:31.533 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:31.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:31 vm10.local ceph-mon[53712]: pgmap v793: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:31 vm11.local ceph-mon[53973]: pgmap v793: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:33 vm10.local ceph-mon[53712]: pgmap v794: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:33 vm11.local ceph-mon[53973]: pgmap v794: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:35 vm10.local ceph-mon[53712]: pgmap v795: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:35 vm11.local ceph-mon[53973]: pgmap v795: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:36.534 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:36.535 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:36.560 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:36.560 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:37 vm11.local ceph-mon[53973]: pgmap v796: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:37 vm10.local ceph-mon[53712]: pgmap v796: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:38 vm10.local ceph-mon[53712]: pgmap v797: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:38 vm11.local ceph-mon[53973]: pgmap v797: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:41.562 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:41.562 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:41 vm10.local ceph-mon[53712]: pgmap v798: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:41.695 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:41.695 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:41 vm11.local ceph-mon[53973]: pgmap v798: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:43 vm10.local ceph-mon[53712]: pgmap v799: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:43 vm11.local ceph-mon[53973]: pgmap v799: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:44.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:44 vm11.local ceph-mon[53973]: pgmap v800: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:44.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:44 vm10.local ceph-mon[53712]: pgmap v800: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:46.697 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:46.697 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:46.723 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:46.724 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:47.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:47 vm11.local ceph-mon[53973]: pgmap v801: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:47 vm10.local ceph-mon[53712]: pgmap v801: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:48 vm10.local ceph-mon[53712]: pgmap v802: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:48 vm11.local ceph-mon[53973]: pgmap v802: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:51 vm10.local ceph-mon[53712]: pgmap v803: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:51.725 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:51.726 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:51.752 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:51.752 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:51 vm11.local ceph-mon[53973]: pgmap v803: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:53.636 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:53 vm10.local ceph-mon[53712]: pgmap v804: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:53 vm11.local ceph-mon[53973]: pgmap v804: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:55 vm10.local ceph-mon[53712]: pgmap v805: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:55 vm11.local ceph-mon[53973]: pgmap v805: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:56.754 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:52:56.754 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:52:56.780 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:52:56.780 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:52:57.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:57 vm11.local ceph-mon[53973]: pgmap v806: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:57 vm10.local ceph-mon[53712]: pgmap v806: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:52:59.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:52:58 vm10.local ceph-mon[53712]: pgmap v807: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:52:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:52:58 vm11.local ceph-mon[53973]: pgmap v807: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:01.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:01 vm11.local ceph-mon[53973]: pgmap v808: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:01.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:01 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:53:01.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:01 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:53:01.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:01 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:53:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:01 vm10.local ceph-mon[53712]: pgmap v808: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:01.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:53:01.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:53:01.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:01 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:53:01.782 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:01.782 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:01.811 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:01.811 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:02.522 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:53:02.522 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:53:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:53:02.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:53:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:03 vm10.local ceph-mon[53712]: pgmap v809: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:03 vm11.local ceph-mon[53973]: pgmap v809: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:05 vm10.local ceph-mon[53712]: pgmap v810: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:05 vm11.local ceph-mon[53973]: pgmap v810: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:06.813 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:06.813 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:06.841 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:06.842 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:07 vm10.local ceph-mon[53712]: pgmap v811: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:07 vm11.local ceph-mon[53973]: pgmap v811: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:08 vm10.local ceph-mon[53712]: pgmap v812: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:08 vm11.local ceph-mon[53973]: pgmap v812: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:11 vm10.local ceph-mon[53712]: pgmap v813: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:11 vm11.local ceph-mon[53973]: pgmap v813: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:11.843 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:11.843 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:11.869 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:11.870 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:13.638 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:13 vm10.local ceph-mon[53712]: pgmap v814: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:13 vm11.local ceph-mon[53973]: pgmap v814: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:15 vm10.local ceph-mon[53712]: pgmap v815: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:15 vm11.local ceph-mon[53973]: pgmap v815: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:16.871 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:16.871 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:16.898 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:16.898 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:17.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:17 vm11.local ceph-mon[53973]: pgmap v816: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:17 vm10.local ceph-mon[53712]: pgmap v816: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:18 vm10.local ceph-mon[53712]: pgmap v817: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:18 vm11.local ceph-mon[53973]: pgmap v817: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:21 vm10.local ceph-mon[53712]: pgmap v818: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:21 vm11.local ceph-mon[53973]: pgmap v818: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:21.900 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:21.900 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:21.925 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:21.926 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:23.639 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:23 vm10.local ceph-mon[53712]: pgmap v819: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:23 vm11.local ceph-mon[53973]: pgmap v819: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:25 vm10.local ceph-mon[53712]: pgmap v820: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:25 vm11.local ceph-mon[53973]: pgmap v820: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:26.927 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:26.928 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:26.954 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:26.954 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:27.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:27 vm11.local ceph-mon[53973]: pgmap v821: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:27 vm10.local ceph-mon[53712]: pgmap v821: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:53:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:53:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:28 vm10.local ceph-mon[53712]: pgmap v822: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:53:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:53:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:28 vm11.local ceph-mon[53973]: pgmap v822: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:31 vm10.local ceph-mon[53712]: pgmap v823: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:31 vm11.local ceph-mon[53973]: pgmap v823: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:31.956 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:31.956 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:31.981 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:31.981 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:33.639 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:33 vm10.local ceph-mon[53712]: pgmap v824: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:33 vm11.local ceph-mon[53973]: pgmap v824: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:35 vm10.local ceph-mon[53712]: pgmap v825: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:35 vm11.local ceph-mon[53973]: pgmap v825: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:36.983 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:36.983 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:37.010 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:37.011 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:37.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:37 vm11.local ceph-mon[53973]: pgmap v826: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:37 vm10.local ceph-mon[53712]: pgmap v826: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:38 vm10.local ceph-mon[53712]: pgmap v827: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:38 vm11.local ceph-mon[53973]: pgmap v827: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:41.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:41 vm10.local ceph-mon[53712]: pgmap v828: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:41 vm11.local ceph-mon[53973]: pgmap v828: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:42.012 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:42.013 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:42.041 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:42.041 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:43.641 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:43 vm10.local ceph-mon[53712]: pgmap v829: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:43 vm11.local ceph-mon[53973]: pgmap v829: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:45 vm10.local ceph-mon[53712]: pgmap v830: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:45 vm11.local ceph-mon[53973]: pgmap v830: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:47.042 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:47.043 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:47.069 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:47.070 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:47.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:47 vm11.local ceph-mon[53973]: pgmap v831: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:47 vm10.local ceph-mon[53712]: pgmap v831: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:48 vm10.local ceph-mon[53712]: pgmap v832: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:48 vm11.local ceph-mon[53973]: pgmap v832: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:51 vm10.local ceph-mon[53712]: pgmap v833: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:51 vm11.local ceph-mon[53973]: pgmap v833: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:52.071 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:52.072 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:52.096 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:52.097 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:53.642 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:53 vm10.local ceph-mon[53712]: pgmap v834: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:53 vm11.local ceph-mon[53973]: pgmap v834: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:55 vm10.local ceph-mon[53712]: pgmap v835: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:55 vm11.local ceph-mon[53973]: pgmap v835: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:57.098 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:53:57.099 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:53:57.125 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:53:57.126 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:53:57.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:57 vm11.local ceph-mon[53973]: pgmap v836: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:57 vm10.local ceph-mon[53712]: pgmap v836: 97 pgs: 97 active+clean; 453 KiB data, 58 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:53:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:53:58 vm10.local ceph-mon[53712]: pgmap v837: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:53:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:53:58 vm11.local ceph-mon[53973]: pgmap v837: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:01.645 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:01 vm11.local ceph-mon[53973]: pgmap v838: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:01.651 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:01 vm10.local ceph-mon[53712]: pgmap v838: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:02.127 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:02.127 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:02.154 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:02.154 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:54:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:54:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:54:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:54:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:54:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:54:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:54:02.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:54:02.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:54:02.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:54:03.643 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:03 vm10.local ceph-mon[53712]: pgmap v839: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:03 vm11.local ceph-mon[53973]: pgmap v839: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:05 vm10.local ceph-mon[53712]: pgmap v840: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:05 vm11.local ceph-mon[53973]: pgmap v840: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:07.156 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:07.156 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:07.181 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:07.182 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:07.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:07 vm11.local ceph-mon[53973]: pgmap v841: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:07 vm10.local ceph-mon[53712]: pgmap v841: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:08 vm10.local ceph-mon[53712]: pgmap v842: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:08 vm11.local ceph-mon[53973]: pgmap v842: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:11 vm10.local ceph-mon[53712]: pgmap v843: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:11 vm11.local ceph-mon[53973]: pgmap v843: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:12.183 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:12.184 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:12.210 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:12.210 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:13.644 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:13 vm10.local ceph-mon[53712]: pgmap v844: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:13 vm11.local ceph-mon[53973]: pgmap v844: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:15 vm10.local ceph-mon[53712]: pgmap v845: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:15 vm11.local ceph-mon[53973]: pgmap v845: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:17.212 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:17.212 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:17.238 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:17.238 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:17.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:17 vm11.local ceph-mon[53973]: pgmap v846: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:17 vm10.local ceph-mon[53712]: pgmap v846: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:18 vm10.local ceph-mon[53712]: pgmap v847: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:18 vm11.local ceph-mon[53973]: pgmap v847: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:21 vm10.local ceph-mon[53712]: pgmap v848: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:21 vm11.local ceph-mon[53973]: pgmap v848: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:22.239 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:22.240 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:22.265 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:22.266 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:23.644 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:23 vm10.local ceph-mon[53712]: pgmap v849: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:23 vm11.local ceph-mon[53973]: pgmap v849: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:25.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:25 vm10.local ceph-mon[53712]: pgmap v850: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:25 vm11.local ceph-mon[53973]: pgmap v850: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:27.267 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:27.268 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:27.293 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:27.293 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:27.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:27 vm11.local ceph-mon[53973]: pgmap v851: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:27 vm10.local ceph-mon[53712]: pgmap v851: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:54:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:54:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:28 vm10.local ceph-mon[53712]: pgmap v852: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:54:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:54:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:28 vm11.local ceph-mon[53973]: pgmap v852: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:31 vm10.local ceph-mon[53712]: pgmap v853: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:31 vm11.local ceph-mon[53973]: pgmap v853: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:32.294 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:32.295 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:32.320 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:32.320 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:33.646 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:33 vm10.local ceph-mon[53712]: pgmap v854: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:33 vm11.local ceph-mon[53973]: pgmap v854: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:35 vm10.local ceph-mon[53712]: pgmap v855: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:35 vm11.local ceph-mon[53973]: pgmap v855: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:37.321 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:37.322 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:37.347 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:37.347 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:37.663 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:37 vm11.local ceph-mon[53973]: pgmap v856: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:37 vm10.local ceph-mon[53712]: pgmap v856: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:38 vm10.local ceph-mon[53712]: pgmap v857: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:38 vm11.local ceph-mon[53973]: pgmap v857: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:41 vm10.local ceph-mon[53712]: pgmap v858: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:41 vm11.local ceph-mon[53973]: pgmap v858: 97 pgs: 97 active+clean; 453 KiB data, 62 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:42.348 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:42.349 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:42.374 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:42.375 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:43.649 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:43 vm10.local ceph-mon[53712]: pgmap v859: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:43 vm11.local ceph-mon[53973]: pgmap v859: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:45 vm10.local ceph-mon[53712]: pgmap v860: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:45 vm11.local ceph-mon[53973]: pgmap v860: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:47.376 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:47.376 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:47.403 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:47.403 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:47 vm10.local ceph-mon[53712]: pgmap v861: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:47 vm11.local ceph-mon[53973]: pgmap v861: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:48 vm10.local ceph-mon[53712]: pgmap v862: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:48 vm11.local ceph-mon[53973]: pgmap v862: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:51 vm10.local ceph-mon[53712]: pgmap v863: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:51 vm11.local ceph-mon[53973]: pgmap v863: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:52.404 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:52.405 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:52.430 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:52.431 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:53.647 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:53 vm10.local ceph-mon[53712]: pgmap v864: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:53 vm11.local ceph-mon[53973]: pgmap v864: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:55 vm10.local ceph-mon[53712]: pgmap v865: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:55 vm11.local ceph-mon[53973]: pgmap v865: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:57.432 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:54:57.433 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:54:57.458 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:54:57.459 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:54:57.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:57 vm11.local ceph-mon[53973]: pgmap v866: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:57 vm10.local ceph-mon[53712]: pgmap v866: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:54:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:54:58 vm10.local ceph-mon[53712]: pgmap v867: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:54:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:54:58 vm11.local ceph-mon[53973]: pgmap v867: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:01 vm10.local ceph-mon[53712]: pgmap v868: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:01.693 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:01 vm11.local ceph-mon[53973]: pgmap v868: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:02.460 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:02.461 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:02.488 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:02.489 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:55:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:55:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:55:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:55:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:55:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:55:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:55:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:55:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:55:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:55:03.649 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:03 vm10.local ceph-mon[53712]: pgmap v869: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:03 vm11.local ceph-mon[53973]: pgmap v869: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:05 vm10.local ceph-mon[53712]: pgmap v870: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:05 vm11.local ceph-mon[53973]: pgmap v870: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:07.490 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:07.491 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:07.517 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:07.517 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:07.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:07 vm11.local ceph-mon[53973]: pgmap v871: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:07 vm10.local ceph-mon[53712]: pgmap v871: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:08 vm10.local ceph-mon[53712]: pgmap v872: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:08 vm11.local ceph-mon[53973]: pgmap v872: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:11 vm10.local ceph-mon[53712]: pgmap v873: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:11 vm11.local ceph-mon[53973]: pgmap v873: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:12.519 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:12.519 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:12.545 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:12.546 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:13 vm10.local ceph-mon[53712]: pgmap v874: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:13 vm11.local ceph-mon[53973]: pgmap v874: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:15 vm10.local ceph-mon[53712]: pgmap v875: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:15 vm11.local ceph-mon[53973]: pgmap v875: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:17.547 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:17.548 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:17.575 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:17.575 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:17 vm10.local ceph-mon[53712]: pgmap v876: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:17 vm11.local ceph-mon[53973]: pgmap v876: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:18 vm10.local ceph-mon[53712]: pgmap v877: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:18 vm11.local ceph-mon[53973]: pgmap v877: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:21 vm10.local ceph-mon[53712]: pgmap v878: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:21 vm11.local ceph-mon[53973]: pgmap v878: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:22.577 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:22.577 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:22.604 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:22.604 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:23.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:23 vm10.local ceph-mon[53712]: pgmap v879: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:23 vm11.local ceph-mon[53973]: pgmap v879: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:25 vm10.local ceph-mon[53712]: pgmap v880: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:25 vm11.local ceph-mon[53973]: pgmap v880: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:27.605 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:27.606 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:27.631 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:27.632 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:27 vm10.local ceph-mon[53712]: pgmap v881: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:27 vm11.local ceph-mon[53973]: pgmap v881: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:55:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:55:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:28 vm10.local ceph-mon[53712]: pgmap v882: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:55:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:55:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:28 vm11.local ceph-mon[53973]: pgmap v882: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:31 vm10.local ceph-mon[53712]: pgmap v883: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:31 vm11.local ceph-mon[53973]: pgmap v883: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:32.633 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:32.633 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:32.659 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:32.659 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:33.655 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:33 vm10.local ceph-mon[53712]: pgmap v884: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:33 vm11.local ceph-mon[53973]: pgmap v884: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:35 vm10.local ceph-mon[53712]: pgmap v885: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:35 vm11.local ceph-mon[53973]: pgmap v885: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:37.660 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:37.661 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:37 vm10.local ceph-mon[53712]: pgmap v886: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:37.687 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:37.687 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:37 vm11.local ceph-mon[53973]: pgmap v886: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:38 vm10.local ceph-mon[53712]: pgmap v887: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:38 vm11.local ceph-mon[53973]: pgmap v887: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:41 vm10.local ceph-mon[53712]: pgmap v888: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:41 vm11.local ceph-mon[53973]: pgmap v888: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:42.688 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:42.689 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:42.714 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:42.714 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:43 vm10.local ceph-mon[53712]: pgmap v889: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:43 vm11.local ceph-mon[53973]: pgmap v889: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:45 vm10.local ceph-mon[53712]: pgmap v890: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:45 vm11.local ceph-mon[53973]: pgmap v890: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:47 vm10.local ceph-mon[53712]: pgmap v891: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:47.715 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:47.716 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:47.743 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:47.743 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:47 vm11.local ceph-mon[53973]: pgmap v891: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:48 vm10.local ceph-mon[53712]: pgmap v892: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:48 vm11.local ceph-mon[53973]: pgmap v892: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:51 vm10.local ceph-mon[53712]: pgmap v893: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:51 vm11.local ceph-mon[53973]: pgmap v893: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:52.745 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:52.745 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:52.770 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:52.770 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:53.656 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:53 vm10.local ceph-mon[53712]: pgmap v894: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:53 vm11.local ceph-mon[53973]: pgmap v894: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:55 vm10.local ceph-mon[53712]: pgmap v895: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:55 vm11.local ceph-mon[53973]: pgmap v895: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:57.771 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:55:57.772 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:55:57.797 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:55:57.797 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:55:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:57 vm11.local ceph-mon[53973]: pgmap v896: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:57 vm10.local ceph-mon[53712]: pgmap v896: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:55:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:55:58 vm10.local ceph-mon[53712]: pgmap v897: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:55:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:55:58 vm11.local ceph-mon[53973]: pgmap v897: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:01 vm10.local ceph-mon[53712]: pgmap v898: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:01 vm11.local ceph-mon[53973]: pgmap v898: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:56:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:56:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:56:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:56:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:56:02.799 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:02.799 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:02.826 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:02.826 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:56:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:56:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:56:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:56:02.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:56:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:03 vm10.local ceph-mon[53712]: pgmap v899: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:03 vm11.local ceph-mon[53973]: pgmap v899: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:05 vm10.local ceph-mon[53712]: pgmap v900: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:05 vm11.local ceph-mon[53973]: pgmap v900: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:07 vm10.local ceph-mon[53712]: pgmap v901: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:07.827 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:07.828 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:07 vm11.local ceph-mon[53973]: pgmap v901: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:07.859 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:07.859 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:08 vm10.local ceph-mon[53712]: pgmap v902: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:08 vm11.local ceph-mon[53973]: pgmap v902: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:11 vm10.local ceph-mon[53712]: pgmap v903: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:11 vm11.local ceph-mon[53973]: pgmap v903: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:12.860 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:12.861 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:12.886 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:12.887 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:13 vm10.local ceph-mon[53712]: pgmap v904: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:13 vm11.local ceph-mon[53973]: pgmap v904: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:15 vm10.local ceph-mon[53712]: pgmap v905: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:15 vm11.local ceph-mon[53973]: pgmap v905: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:17 vm10.local ceph-mon[53712]: pgmap v906: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:17 vm11.local ceph-mon[53973]: pgmap v906: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:17.888 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:17.889 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:17.913 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:17.914 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:18 vm10.local ceph-mon[53712]: pgmap v907: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:18 vm11.local ceph-mon[53973]: pgmap v907: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:21 vm10.local ceph-mon[53712]: pgmap v908: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:21 vm11.local ceph-mon[53973]: pgmap v908: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:22.915 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:22.916 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:22.940 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:22.941 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:23.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:23 vm10.local ceph-mon[53712]: pgmap v909: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:23 vm11.local ceph-mon[53973]: pgmap v909: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:25 vm10.local ceph-mon[53712]: pgmap v910: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:25 vm11.local ceph-mon[53973]: pgmap v910: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:27.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:27 vm10.local ceph-mon[53712]: pgmap v911: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:27 vm11.local ceph-mon[53973]: pgmap v911: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:27.942 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:27.943 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:27.969 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:27.970 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:56:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:56:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:28 vm10.local ceph-mon[53712]: pgmap v912: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:56:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:56:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:28 vm11.local ceph-mon[53973]: pgmap v912: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:31 vm10.local ceph-mon[53712]: pgmap v913: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:31 vm11.local ceph-mon[53973]: pgmap v913: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:32.971 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:32.972 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:32.997 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:32.998 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:33 vm10.local ceph-mon[53712]: pgmap v914: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:33 vm11.local ceph-mon[53973]: pgmap v914: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:35 vm10.local ceph-mon[53712]: pgmap v915: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:35 vm11.local ceph-mon[53973]: pgmap v915: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:37 vm10.local ceph-mon[53712]: pgmap v916: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:37 vm11.local ceph-mon[53973]: pgmap v916: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:37.999 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:37.999 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:38.025 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:38.026 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:38 vm10.local ceph-mon[53712]: pgmap v917: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:38 vm11.local ceph-mon[53973]: pgmap v917: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:41 vm10.local ceph-mon[53712]: pgmap v918: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:41 vm11.local ceph-mon[53973]: pgmap v918: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:43.027 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:43.027 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:43.052 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:43.052 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:43 vm10.local ceph-mon[53712]: pgmap v919: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:43 vm11.local ceph-mon[53973]: pgmap v919: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:45 vm10.local ceph-mon[53712]: pgmap v920: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:45 vm11.local ceph-mon[53973]: pgmap v920: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:47 vm10.local ceph-mon[53712]: pgmap v921: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:47 vm11.local ceph-mon[53973]: pgmap v921: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:48.053 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:48.054 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:48.079 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:48.079 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:48 vm10.local ceph-mon[53712]: pgmap v922: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:48 vm11.local ceph-mon[53973]: pgmap v922: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:51 vm10.local ceph-mon[53712]: pgmap v923: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:51 vm11.local ceph-mon[53973]: pgmap v923: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:53.081 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:53.081 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:53.106 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:53.107 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:53 vm10.local ceph-mon[53712]: pgmap v924: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:53 vm11.local ceph-mon[53973]: pgmap v924: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:55 vm10.local ceph-mon[53712]: pgmap v925: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:55 vm11.local ceph-mon[53973]: pgmap v925: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:57 vm10.local ceph-mon[53712]: pgmap v926: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:57 vm11.local ceph-mon[53973]: pgmap v926: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:56:58.108 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:56:58.108 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:56:58.135 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:56:58.136 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:56:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:56:58 vm10.local ceph-mon[53712]: pgmap v927: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:56:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:56:58 vm11.local ceph-mon[53973]: pgmap v927: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:01 vm10.local ceph-mon[53712]: pgmap v928: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:01 vm11.local ceph-mon[53973]: pgmap v928: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:02.418 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:57:02.418 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:57:02.418 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:57:02.418 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:02 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:57:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:57:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:57:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:57:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:02 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:57:03.137 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:03.138 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:03.163 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:03.163 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:03 vm10.local ceph-mon[53712]: pgmap v929: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:57:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:57:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:03 vm11.local ceph-mon[53973]: pgmap v929: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:57:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:57:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:05 vm11.local ceph-mon[53973]: pgmap v930: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:05 vm10.local ceph-mon[53712]: pgmap v930: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:07.502 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:07 vm10.local ceph-mon[53712]: pgmap v931: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:07 vm11.local ceph-mon[53973]: pgmap v931: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:08.165 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:08.166 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:08.191 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:08.191 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:08 vm10.local ceph-mon[53712]: pgmap v932: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:08 vm11.local ceph-mon[53973]: pgmap v932: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:11 vm10.local ceph-mon[53712]: pgmap v933: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:11 vm11.local ceph-mon[53973]: pgmap v933: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:13.193 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:13.193 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:13.218 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:13.219 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:13.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:13 vm10.local ceph-mon[53712]: pgmap v934: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:13 vm11.local ceph-mon[53973]: pgmap v934: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:15 vm10.local ceph-mon[53712]: pgmap v935: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:15 vm11.local ceph-mon[53973]: pgmap v935: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:17 vm10.local ceph-mon[53712]: pgmap v936: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:17 vm11.local ceph-mon[53973]: pgmap v936: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:18.220 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:18.220 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:18.245 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:18.246 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:18 vm10.local ceph-mon[53712]: pgmap v937: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:18 vm11.local ceph-mon[53973]: pgmap v937: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:21.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:21 vm10.local ceph-mon[53712]: pgmap v938: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:21 vm11.local ceph-mon[53973]: pgmap v938: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:23.247 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:23.248 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:23.275 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:23.275 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:23.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:23 vm10.local ceph-mon[53712]: pgmap v939: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:23 vm11.local ceph-mon[53973]: pgmap v939: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:25 vm10.local ceph-mon[53712]: pgmap v940: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:25 vm11.local ceph-mon[53973]: pgmap v940: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:27 vm11.local ceph-mon[53973]: pgmap v941: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:27 vm10.local ceph-mon[53712]: pgmap v941: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:28.277 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:28.277 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:28.303 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:28.304 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:57:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:57:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:28 vm10.local ceph-mon[53712]: pgmap v942: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:57:29.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:57:29.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:28 vm11.local ceph-mon[53973]: pgmap v942: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:31 vm10.local ceph-mon[53712]: pgmap v943: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:31 vm11.local ceph-mon[53973]: pgmap v943: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:33.305 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:33.306 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:33.332 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:33.333 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:33 vm10.local ceph-mon[53712]: pgmap v944: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:33 vm11.local ceph-mon[53973]: pgmap v944: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:35.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:35 vm10.local ceph-mon[53712]: pgmap v945: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:35 vm11.local ceph-mon[53973]: pgmap v945: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:37 vm11.local ceph-mon[53973]: pgmap v946: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:37 vm10.local ceph-mon[53712]: pgmap v946: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:38.334 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:38.335 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:38.360 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:38.361 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:38 vm10.local ceph-mon[53712]: pgmap v947: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:38 vm11.local ceph-mon[53973]: pgmap v947: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:41.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:41 vm10.local ceph-mon[53712]: pgmap v948: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:41 vm11.local ceph-mon[53973]: pgmap v948: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:43.363 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:43.363 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:43.391 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:43.392 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:43 vm10.local ceph-mon[53712]: pgmap v949: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:43 vm11.local ceph-mon[53973]: pgmap v949: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:45 vm11.local ceph-mon[53973]: pgmap v950: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:45 vm10.local ceph-mon[53712]: pgmap v950: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:47 vm11.local ceph-mon[53973]: pgmap v951: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:47 vm10.local ceph-mon[53712]: pgmap v951: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:48.393 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:48.394 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:48.419 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:48.420 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:48 vm10.local ceph-mon[53712]: pgmap v952: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:48 vm11.local ceph-mon[53973]: pgmap v952: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:51.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:51 vm10.local ceph-mon[53712]: pgmap v953: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:51 vm11.local ceph-mon[53973]: pgmap v953: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:53.421 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:53.422 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:53.447 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:53.448 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:53 vm10.local ceph-mon[53712]: pgmap v954: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:53 vm11.local ceph-mon[53973]: pgmap v954: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:55 vm11.local ceph-mon[53973]: pgmap v955: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:55 vm10.local ceph-mon[53712]: pgmap v955: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:57 vm11.local ceph-mon[53973]: pgmap v956: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:57 vm10.local ceph-mon[53712]: pgmap v956: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:57:58.449 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:57:58.450 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:57:58.475 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:57:58.476 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:57:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:57:59 vm11.local ceph-mon[53973]: pgmap v957: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:57:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:57:59 vm10.local ceph-mon[53712]: pgmap v957: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:01 vm10.local ceph-mon[53712]: pgmap v958: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:01 vm11.local ceph-mon[53973]: pgmap v958: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:03.477 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:03.478 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:03.508 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:03.508 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: pgmap v959: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: pgmap v959: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:03.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:58:03.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:58:03.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:58:03.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:03.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:58:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:05 vm11.local ceph-mon[53973]: pgmap v960: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:05.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:05 vm10.local ceph-mon[53712]: pgmap v960: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:07 vm11.local ceph-mon[53973]: pgmap v961: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:07.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:07 vm10.local ceph-mon[53712]: pgmap v961: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:08.510 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:08.510 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:08.539 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:08.540 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:09 vm10.local ceph-mon[53712]: pgmap v962: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:09 vm11.local ceph-mon[53973]: pgmap v962: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:11 vm10.local ceph-mon[53712]: pgmap v963: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:11 vm11.local ceph-mon[53973]: pgmap v963: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:13.542 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:13.542 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:13.568 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:13.568 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:13 vm11.local ceph-mon[53973]: pgmap v964: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:13 vm10.local ceph-mon[53712]: pgmap v964: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:15 vm11.local ceph-mon[53973]: pgmap v965: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:15 vm10.local ceph-mon[53712]: pgmap v965: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:17 vm11.local ceph-mon[53973]: pgmap v966: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:17 vm10.local ceph-mon[53712]: pgmap v966: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:18.569 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:18.570 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:18.597 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:18.597 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:19 vm10.local ceph-mon[53712]: pgmap v967: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:19 vm11.local ceph-mon[53973]: pgmap v967: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:21 vm10.local ceph-mon[53712]: pgmap v968: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:21 vm11.local ceph-mon[53973]: pgmap v968: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:23.598 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:23.599 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:23.626 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:23.627 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:23 vm11.local ceph-mon[53973]: pgmap v969: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:23 vm10.local ceph-mon[53712]: pgmap v969: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:25 vm11.local ceph-mon[53973]: pgmap v970: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:25 vm10.local ceph-mon[53712]: pgmap v970: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:27 vm11.local ceph-mon[53973]: pgmap v971: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:27 vm10.local ceph-mon[53712]: pgmap v971: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:28.628 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:28.629 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:28.657 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:28.658 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:58:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:58:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:29 vm10.local ceph-mon[53712]: pgmap v972: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:58:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:58:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:29 vm11.local ceph-mon[53973]: pgmap v972: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:31 vm10.local ceph-mon[53712]: pgmap v973: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:31 vm11.local ceph-mon[53973]: pgmap v973: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:33.659 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:33.660 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:33.687 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:33.688 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:33 vm11.local ceph-mon[53973]: pgmap v974: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:33 vm10.local ceph-mon[53712]: pgmap v974: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:35 vm11.local ceph-mon[53973]: pgmap v975: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:35 vm10.local ceph-mon[53712]: pgmap v975: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:37 vm11.local ceph-mon[53973]: pgmap v976: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:37 vm10.local ceph-mon[53712]: pgmap v976: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:38.689 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:38.690 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:38.715 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:38.716 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:39 vm10.local ceph-mon[53712]: pgmap v977: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:39 vm11.local ceph-mon[53973]: pgmap v977: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:41 vm10.local ceph-mon[53712]: pgmap v978: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:41 vm11.local ceph-mon[53973]: pgmap v978: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:43.717 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:43.718 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:43.743 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:43.744 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:43 vm11.local ceph-mon[53973]: pgmap v979: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:43 vm10.local ceph-mon[53712]: pgmap v979: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:45 vm11.local ceph-mon[53973]: pgmap v980: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:45.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:45 vm10.local ceph-mon[53712]: pgmap v980: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:47.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:47 vm11.local ceph-mon[53973]: pgmap v981: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:47 vm10.local ceph-mon[53712]: pgmap v981: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:48.745 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:48.746 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:48.771 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:48.772 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:49.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:49 vm10.local ceph-mon[53712]: pgmap v982: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:49 vm11.local ceph-mon[53973]: pgmap v982: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:51 vm11.local ceph-mon[53973]: pgmap v983: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:51.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:51 vm10.local ceph-mon[53712]: pgmap v983: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:53.773 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:53.774 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:53.802 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:53.803 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:53 vm11.local ceph-mon[53973]: pgmap v984: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:53.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:53 vm10.local ceph-mon[53712]: pgmap v984: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:55 vm11.local ceph-mon[53973]: pgmap v985: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:55.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:55 vm10.local ceph-mon[53712]: pgmap v985: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:57 vm11.local ceph-mon[53973]: pgmap v986: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:57 vm10.local ceph-mon[53712]: pgmap v986: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:58:58.805 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:58:58.805 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:58:58.833 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:58:58.834 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:58:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:58:59 vm10.local ceph-mon[53712]: pgmap v987: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:58:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:58:59 vm11.local ceph-mon[53973]: pgmap v987: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:01 vm11.local ceph-mon[53973]: pgmap v988: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:01 vm10.local ceph-mon[53712]: pgmap v988: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:03.510 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:03 vm11.local ceph-mon[53973]: pgmap v989: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:03.510 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:59:03.510 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:59:03.510 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:03 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:59:03.546 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:03 vm10.local ceph-mon[53712]: pgmap v989: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:03.546 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T17:59:03.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T17:59:03.547 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:03 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T17:59:03.839 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:03.842 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:03.893 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:03.893 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:05 vm10.local ceph-mon[53712]: pgmap v990: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:05 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:59:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:05 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:59:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:05 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:59:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:05 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:59:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:05 vm11.local ceph-mon[53973]: pgmap v990: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:05 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:59:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:05 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T17:59:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:05 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:59:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:05 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T17:59:07.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:07 vm11.local ceph-mon[53973]: pgmap v991: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:07.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:07 vm10.local ceph-mon[53712]: pgmap v991: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:08.895 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:08.896 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:08.926 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:08.927 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:09 vm10.local ceph-mon[53712]: pgmap v992: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:09 vm11.local ceph-mon[53973]: pgmap v992: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:11 vm11.local ceph-mon[53973]: pgmap v993: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:11.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:11 vm10.local ceph-mon[53712]: pgmap v993: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:13 vm11.local ceph-mon[53973]: pgmap v994: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:13.928 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:13.929 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:13 vm10.local ceph-mon[53712]: pgmap v994: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:13.997 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:13.997 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:15 vm10.local ceph-mon[53712]: pgmap v995: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:15 vm11.local ceph-mon[53973]: pgmap v995: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:17 vm11.local ceph-mon[53973]: pgmap v996: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:17 vm10.local ceph-mon[53712]: pgmap v996: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:18.999 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:18.999 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:19.029 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:19.030 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:19 vm10.local ceph-mon[53712]: pgmap v997: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:19 vm11.local ceph-mon[53973]: pgmap v997: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:21 vm11.local ceph-mon[53973]: pgmap v998: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:21 vm10.local ceph-mon[53712]: pgmap v998: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:23 vm11.local ceph-mon[53973]: pgmap v999: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:23 vm10.local ceph-mon[53712]: pgmap v999: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:24.031 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:24.031 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:24.061 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:24.061 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:25 vm11.local ceph-mon[53973]: pgmap v1000: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:25 vm10.local ceph-mon[53712]: pgmap v1000: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:27 vm11.local ceph-mon[53973]: pgmap v1001: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:27 vm10.local ceph-mon[53712]: pgmap v1001: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:29.062 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:29.063 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:29.088 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:29.088 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:59:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:59:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:29 vm10.local ceph-mon[53712]: pgmap v1002: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T17:59:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T17:59:29.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:29 vm11.local ceph-mon[53973]: pgmap v1002: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:31 vm11.local ceph-mon[53973]: pgmap v1003: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:31 vm10.local ceph-mon[53712]: pgmap v1003: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:33 vm11.local ceph-mon[53973]: pgmap v1004: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:33 vm10.local ceph-mon[53712]: pgmap v1004: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:34.090 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:34.090 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:34.117 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:34.117 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:35 vm11.local ceph-mon[53973]: pgmap v1005: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:35.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:35 vm10.local ceph-mon[53712]: pgmap v1005: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:37 vm11.local ceph-mon[53973]: pgmap v1006: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:37.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:37 vm10.local ceph-mon[53712]: pgmap v1006: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:39.119 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:39.119 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:39.144 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:39.145 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:39 vm10.local ceph-mon[53712]: pgmap v1007: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:39 vm11.local ceph-mon[53973]: pgmap v1007: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:41 vm11.local ceph-mon[53973]: pgmap v1008: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:41 vm10.local ceph-mon[53712]: pgmap v1008: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:43 vm11.local ceph-mon[53973]: pgmap v1009: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:43 vm10.local ceph-mon[53712]: pgmap v1009: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:44.146 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:44.146 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:44.172 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:44.172 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:45.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:45 vm11.local ceph-mon[53973]: pgmap v1010: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:45 vm10.local ceph-mon[53712]: pgmap v1010: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:47 vm11.local ceph-mon[53973]: pgmap v1011: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:47 vm10.local ceph-mon[53712]: pgmap v1011: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:49.174 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:49.174 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:49 vm10.local ceph-mon[53712]: pgmap v1012: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:49.199 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:49.200 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:49 vm11.local ceph-mon[53973]: pgmap v1012: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:51 vm11.local ceph-mon[53973]: pgmap v1013: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:51 vm10.local ceph-mon[53712]: pgmap v1013: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:53 vm11.local ceph-mon[53973]: pgmap v1014: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:53 vm10.local ceph-mon[53712]: pgmap v1014: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:54.201 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:54.202 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:54.228 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:54.229 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:55 vm11.local ceph-mon[53973]: pgmap v1015: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:55 vm10.local ceph-mon[53712]: pgmap v1015: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:57 vm11.local ceph-mon[53973]: pgmap v1016: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:57 vm10.local ceph-mon[53712]: pgmap v1016: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T17:59:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 17:59:59 vm10.local ceph-mon[53712]: pgmap v1017: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T17:59:59.230 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T17:59:59.231 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T17:59:59.256 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T17:59:59.257 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T17:59:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 17:59:59 vm11.local ceph-mon[53973]: pgmap v1017: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:00 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T18:00:00.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:00 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T18:00:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:01 vm11.local ceph-mon[53973]: pgmap v1018: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:01 vm10.local ceph-mon[53712]: pgmap v1018: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:03 vm11.local ceph-mon[53973]: pgmap v1019: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:03 vm10.local ceph-mon[53712]: pgmap v1019: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:04.258 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:04.259 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:04.363 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:04.364 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:05.588 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:05 vm11.local ceph-mon[53973]: pgmap v1020: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:05.588 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:05 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:00:05.588 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:05 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:00:05.588 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:05 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:00:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:05 vm10.local ceph-mon[53712]: pgmap v1020: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:05 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:00:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:05 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:00:05.658 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:05 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:00:07.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:06 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:07.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:06 vm10.local ceph-mon[53712]: pgmap v1021: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:07.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:06 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:07.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:06 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:07.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:06 vm11.local ceph-mon[53973]: pgmap v1021: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:07.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:06 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:08.736 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:08.736 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:00:08.736 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:08.736 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:08.793 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:08.793 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:00:08.793 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:08.793 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:00:09.365 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:09.366 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:09.391 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:09.392 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:09 vm11.local ceph-mon[53973]: pgmap v1022: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:09 vm10.local ceph-mon[53712]: pgmap v1022: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:11 vm11.local ceph-mon[53973]: pgmap v1023: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:11 vm10.local ceph-mon[53712]: pgmap v1023: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:13 vm11.local ceph-mon[53973]: pgmap v1024: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:13 vm10.local ceph-mon[53712]: pgmap v1024: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:14.393 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:14.394 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:14.422 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:14.423 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:15 vm11.local ceph-mon[53973]: pgmap v1025: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:15.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:15 vm10.local ceph-mon[53712]: pgmap v1025: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:17 vm11.local ceph-mon[53973]: pgmap v1026: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:17.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:17 vm10.local ceph-mon[53712]: pgmap v1026: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:19 vm10.local ceph-mon[53712]: pgmap v1027: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:19.194 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:19 vm11.local ceph-mon[53973]: pgmap v1027: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:19.425 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:19.425 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:19.460 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:19.461 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:21 vm11.local ceph-mon[53973]: pgmap v1028: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:21 vm10.local ceph-mon[53712]: pgmap v1028: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:23 vm11.local ceph-mon[53973]: pgmap v1029: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:23 vm10.local ceph-mon[53712]: pgmap v1029: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:24.463 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:24.464 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:24.495 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:24.495 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:25 vm11.local ceph-mon[53973]: pgmap v1030: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:25 vm10.local ceph-mon[53712]: pgmap v1030: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:27 vm11.local ceph-mon[53973]: pgmap v1031: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:27 vm10.local ceph-mon[53712]: pgmap v1031: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:00:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:00:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:29 vm10.local ceph-mon[53712]: pgmap v1032: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:00:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:00:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:29 vm11.local ceph-mon[53973]: pgmap v1032: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:29.497 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:29.498 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:29.523 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:29.524 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:31 vm11.local ceph-mon[53973]: pgmap v1033: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:31 vm10.local ceph-mon[53712]: pgmap v1033: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:33 vm11.local ceph-mon[53973]: pgmap v1034: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:33.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:33 vm10.local ceph-mon[53712]: pgmap v1034: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:34.526 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:34.526 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:34.551 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:34.552 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:35 vm11.local ceph-mon[53973]: pgmap v1035: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:35 vm10.local ceph-mon[53712]: pgmap v1035: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:37 vm11.local ceph-mon[53973]: pgmap v1036: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:37 vm10.local ceph-mon[53712]: pgmap v1036: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:39 vm10.local ceph-mon[53712]: pgmap v1037: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:39 vm11.local ceph-mon[53973]: pgmap v1037: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:39.553 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:39.554 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:39.581 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:39.582 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:41 vm11.local ceph-mon[53973]: pgmap v1038: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:41 vm10.local ceph-mon[53712]: pgmap v1038: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:43 vm11.local ceph-mon[53973]: pgmap v1039: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:43.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:43 vm10.local ceph-mon[53712]: pgmap v1039: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:44.583 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:44.584 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:44.610 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:44.611 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:45 vm11.local ceph-mon[53973]: pgmap v1040: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:45 vm10.local ceph-mon[53712]: pgmap v1040: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:47 vm11.local ceph-mon[53973]: pgmap v1041: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:47 vm10.local ceph-mon[53712]: pgmap v1041: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:49 vm10.local ceph-mon[53712]: pgmap v1042: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:49 vm11.local ceph-mon[53973]: pgmap v1042: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:49.612 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:49.613 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:49.639 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:49.639 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:51 vm11.local ceph-mon[53973]: pgmap v1043: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:51 vm10.local ceph-mon[53712]: pgmap v1043: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:53 vm11.local ceph-mon[53973]: pgmap v1044: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:53 vm10.local ceph-mon[53712]: pgmap v1044: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:54.641 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:54.641 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:54.667 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:54.668 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:00:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:55 vm11.local ceph-mon[53973]: pgmap v1045: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:55 vm10.local ceph-mon[53712]: pgmap v1045: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:57 vm11.local ceph-mon[53973]: pgmap v1046: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:57 vm10.local ceph-mon[53712]: pgmap v1046: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:00:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:00:59 vm11.local ceph-mon[53973]: pgmap v1047: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:00:59 vm10.local ceph-mon[53712]: pgmap v1047: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:00:59.669 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:00:59.669 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:00:59.696 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:00:59.697 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:01 vm11.local ceph-mon[53973]: pgmap v1048: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:01.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:01 vm10.local ceph-mon[53712]: pgmap v1048: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:03 vm11.local ceph-mon[53973]: pgmap v1049: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:03.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:03 vm10.local ceph-mon[53712]: pgmap v1049: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:04.699 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:04.699 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:04.727 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:04.727 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:05 vm11.local ceph-mon[53973]: pgmap v1050: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:05 vm10.local ceph-mon[53712]: pgmap v1050: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:06 vm11.local ceph-mon[53973]: pgmap v1051: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:06 vm10.local ceph-mon[53712]: pgmap v1051: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:08.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:07 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:01:08.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:07 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:01:08.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:07 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:01:08.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:07 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:01:08.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:07 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:01:08.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:07 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:01:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:01:09.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:01:09.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:08 vm10.local ceph-mon[53712]: pgmap v1052: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:01:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:01:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:08 vm11.local ceph-mon[53973]: pgmap v1052: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:09.728 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:09.729 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:09.759 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:09.759 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:11 vm11.local ceph-mon[53973]: pgmap v1053: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:11 vm10.local ceph-mon[53712]: pgmap v1053: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:13 vm11.local ceph-mon[53973]: pgmap v1054: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:13 vm10.local ceph-mon[53712]: pgmap v1054: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:14.761 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:14.761 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:14.789 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:14.790 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:15 vm11.local ceph-mon[53973]: pgmap v1055: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:15 vm10.local ceph-mon[53712]: pgmap v1055: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:17 vm11.local ceph-mon[53973]: pgmap v1056: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:17.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:17 vm10.local ceph-mon[53712]: pgmap v1056: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:19.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:19 vm10.local ceph-mon[53712]: pgmap v1057: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:19 vm11.local ceph-mon[53973]: pgmap v1057: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:19.791 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:19.792 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:19.817 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:19.818 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:21 vm11.local ceph-mon[53973]: pgmap v1058: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:21 vm10.local ceph-mon[53712]: pgmap v1058: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:23 vm11.local ceph-mon[53973]: pgmap v1059: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:23 vm10.local ceph-mon[53712]: pgmap v1059: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:24.819 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:24.820 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:24.847 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:24.847 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:25 vm11.local ceph-mon[53973]: pgmap v1060: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:25 vm10.local ceph-mon[53712]: pgmap v1060: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:27 vm11.local ceph-mon[53973]: pgmap v1061: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:27 vm10.local ceph-mon[53712]: pgmap v1061: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:01:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:01:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:29 vm11.local ceph-mon[53973]: pgmap v1062: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:01:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:01:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:29 vm10.local ceph-mon[53712]: pgmap v1062: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:29.849 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:29.849 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:29.875 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:29.875 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:31 vm11.local ceph-mon[53973]: pgmap v1063: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:31 vm10.local ceph-mon[53712]: pgmap v1063: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:33 vm11.local ceph-mon[53973]: pgmap v1064: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:33 vm10.local ceph-mon[53712]: pgmap v1064: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:34.877 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:34.878 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:34.903 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:34.904 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:35 vm11.local ceph-mon[53973]: pgmap v1065: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:35 vm10.local ceph-mon[53712]: pgmap v1065: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:37 vm11.local ceph-mon[53973]: pgmap v1066: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:37 vm10.local ceph-mon[53712]: pgmap v1066: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:39 vm11.local ceph-mon[53973]: pgmap v1067: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:39 vm10.local ceph-mon[53712]: pgmap v1067: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:39.905 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:39.906 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:39.932 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:39.933 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:41 vm11.local ceph-mon[53973]: pgmap v1068: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:41 vm10.local ceph-mon[53712]: pgmap v1068: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:43 vm11.local ceph-mon[53973]: pgmap v1069: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:43 vm10.local ceph-mon[53712]: pgmap v1069: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:44.934 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:44.935 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:44.966 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:44.966 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:45 vm11.local ceph-mon[53973]: pgmap v1070: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:45 vm10.local ceph-mon[53712]: pgmap v1070: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:47 vm11.local ceph-mon[53973]: pgmap v1071: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:47 vm10.local ceph-mon[53712]: pgmap v1071: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:49 vm11.local ceph-mon[53973]: pgmap v1072: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:49.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:49 vm10.local ceph-mon[53712]: pgmap v1072: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:49.968 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:49.968 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:49.994 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:49.994 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:51 vm11.local ceph-mon[53973]: pgmap v1073: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:51 vm10.local ceph-mon[53712]: pgmap v1073: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:53 vm11.local ceph-mon[53973]: pgmap v1074: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:53 vm10.local ceph-mon[53712]: pgmap v1074: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:54.996 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:01:54.996 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:01:55.028 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:01:55.028 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:01:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:54 vm11.local ceph-mon[53973]: pgmap v1075: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:54 vm10.local ceph-mon[53712]: pgmap v1075: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:57 vm11.local ceph-mon[53973]: pgmap v1076: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:57 vm10.local ceph-mon[53712]: pgmap v1076: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:01:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:01:59 vm11.local ceph-mon[53973]: pgmap v1077: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:01:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:01:59 vm10.local ceph-mon[53712]: pgmap v1077: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:00.030 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:00.030 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:00.055 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:00.056 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:01 vm11.local ceph-mon[53973]: pgmap v1078: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:01 vm10.local ceph-mon[53712]: pgmap v1078: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:03 vm11.local ceph-mon[53973]: pgmap v1079: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:03 vm10.local ceph-mon[53712]: pgmap v1079: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:05.057 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:05.057 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:05.082 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:05.083 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:05 vm11.local ceph-mon[53973]: pgmap v1080: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:05 vm10.local ceph-mon[53712]: pgmap v1080: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:07 vm11.local ceph-mon[53973]: pgmap v1081: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:07 vm10.local ceph-mon[53712]: pgmap v1081: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:08.801 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:02:08.801 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:02:08.801 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:02:08.801 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:02:08.801 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:08 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:02:08.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:02:08.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:02:08.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:02:08.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:02:08.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:08 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:02:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:09 vm11.local ceph-mon[53973]: pgmap v1082: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:09 vm10.local ceph-mon[53712]: pgmap v1082: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:10.084 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:10.085 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:10.113 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:10.114 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:11 vm11.local ceph-mon[53973]: pgmap v1083: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:11 vm10.local ceph-mon[53712]: pgmap v1083: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:13 vm11.local ceph-mon[53973]: pgmap v1084: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:13 vm10.local ceph-mon[53712]: pgmap v1084: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:15.115 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:15.116 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:15.145 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:15.145 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:15 vm11.local ceph-mon[53973]: pgmap v1085: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:15 vm10.local ceph-mon[53712]: pgmap v1085: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:17 vm11.local ceph-mon[53973]: pgmap v1086: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:17 vm10.local ceph-mon[53712]: pgmap v1086: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:19 vm11.local ceph-mon[53973]: pgmap v1087: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:19.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:19 vm10.local ceph-mon[53712]: pgmap v1087: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:20.147 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:20.147 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:20.173 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:20.174 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:21 vm11.local ceph-mon[53973]: pgmap v1088: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:21.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:21 vm10.local ceph-mon[53712]: pgmap v1088: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:23.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:23 vm11.local ceph-mon[53973]: pgmap v1089: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:23.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:23 vm10.local ceph-mon[53712]: pgmap v1089: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:25.175 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:25.176 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:25.203 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:25.204 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:25.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:25 vm11.local ceph-mon[53973]: pgmap v1090: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:25.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:25 vm10.local ceph-mon[53712]: pgmap v1090: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:27 vm11.local ceph-mon[53973]: pgmap v1091: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:27 vm10.local ceph-mon[53712]: pgmap v1091: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:02:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:02:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:29 vm10.local ceph-mon[53712]: pgmap v1092: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:02:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:02:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:29 vm11.local ceph-mon[53973]: pgmap v1092: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:30.205 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:30.205 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:30.231 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:30.232 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:31 vm11.local ceph-mon[53973]: pgmap v1093: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:31 vm10.local ceph-mon[53712]: pgmap v1093: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:33 vm11.local ceph-mon[53973]: pgmap v1094: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:33 vm10.local ceph-mon[53712]: pgmap v1094: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:35.233 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:35.234 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:35.260 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:35.260 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:35 vm11.local ceph-mon[53973]: pgmap v1095: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:35 vm10.local ceph-mon[53712]: pgmap v1095: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:37.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:36 vm11.local ceph-mon[53973]: pgmap v1096: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:37.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:36 vm10.local ceph-mon[53712]: pgmap v1096: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:39 vm10.local ceph-mon[53712]: pgmap v1097: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:39 vm11.local ceph-mon[53973]: pgmap v1097: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:40.261 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:40.262 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:40.289 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:40.290 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:41 vm11.local ceph-mon[53973]: pgmap v1098: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:41 vm10.local ceph-mon[53712]: pgmap v1098: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:43 vm11.local ceph-mon[53973]: pgmap v1099: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:43 vm10.local ceph-mon[53712]: pgmap v1099: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:45.291 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:45.292 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:45.319 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:45.319 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:45 vm11.local ceph-mon[53973]: pgmap v1100: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:45 vm10.local ceph-mon[53712]: pgmap v1100: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:47 vm11.local ceph-mon[53973]: pgmap v1101: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:47 vm10.local ceph-mon[53712]: pgmap v1101: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:49.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:49 vm10.local ceph-mon[53712]: pgmap v1102: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:49 vm11.local ceph-mon[53973]: pgmap v1102: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:50.320 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:50.321 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:50.346 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:50.347 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:51 vm11.local ceph-mon[53973]: pgmap v1103: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:51.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:51 vm10.local ceph-mon[53712]: pgmap v1103: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:53 vm11.local ceph-mon[53973]: pgmap v1104: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:53 vm10.local ceph-mon[53712]: pgmap v1104: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:55.349 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:02:55.349 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:02:55.379 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:02:55.380 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:02:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:55 vm11.local ceph-mon[53973]: pgmap v1105: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:55 vm10.local ceph-mon[53712]: pgmap v1105: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:57 vm11.local ceph-mon[53973]: pgmap v1106: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:57 vm10.local ceph-mon[53712]: pgmap v1106: 97 pgs: 97 active+clean; 453 KiB data, 66 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:02:59.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:02:59 vm10.local ceph-mon[53712]: pgmap v1107: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:02:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:02:59 vm11.local ceph-mon[53973]: pgmap v1107: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:00.381 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:00.382 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:00.414 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:00.415 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:01 vm11.local ceph-mon[53973]: pgmap v1108: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:01 vm10.local ceph-mon[53712]: pgmap v1108: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:03 vm11.local ceph-mon[53973]: pgmap v1109: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:03 vm10.local ceph-mon[53712]: pgmap v1109: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:05.416 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:05.417 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:05.442 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:05.442 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:05 vm11.local ceph-mon[53973]: pgmap v1110: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:05 vm10.local ceph-mon[53712]: pgmap v1110: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:07 vm11.local ceph-mon[53973]: pgmap v1111: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:07 vm10.local ceph-mon[53712]: pgmap v1111: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:03:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:03:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:03:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:09 vm10.local ceph-mon[53712]: pgmap v1112: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:03:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:03:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:03:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:03:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:03:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:09 vm11.local ceph-mon[53973]: pgmap v1112: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:03:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:03:10.444 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:10.444 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:10.469 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:10.469 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:11 vm11.local ceph-mon[53973]: pgmap v1113: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:11 vm10.local ceph-mon[53712]: pgmap v1113: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:13 vm11.local ceph-mon[53973]: pgmap v1114: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:13 vm10.local ceph-mon[53712]: pgmap v1114: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:15.471 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:15.471 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:15.501 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:15.502 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:15 vm11.local ceph-mon[53973]: pgmap v1115: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:15 vm10.local ceph-mon[53712]: pgmap v1115: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:17 vm11.local ceph-mon[53973]: pgmap v1116: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:17 vm10.local ceph-mon[53712]: pgmap v1116: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:19.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:19 vm10.local ceph-mon[53712]: pgmap v1117: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:19 vm11.local ceph-mon[53973]: pgmap v1117: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:20.503 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:20.504 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:20.533 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:20.533 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:21 vm11.local ceph-mon[53973]: pgmap v1118: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:21 vm10.local ceph-mon[53712]: pgmap v1118: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:23 vm11.local ceph-mon[53973]: pgmap v1119: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:23.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:23 vm10.local ceph-mon[53712]: pgmap v1119: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:25.534 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:25.535 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:25.576 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:25.576 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:25 vm10.local ceph-mon[53712]: pgmap v1120: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:26.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:25 vm11.local ceph-mon[53973]: pgmap v1120: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:27 vm10.local ceph-mon[53712]: pgmap v1121: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:27 vm11.local ceph-mon[53973]: pgmap v1121: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:03:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:03:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:29 vm11.local ceph-mon[53973]: pgmap v1122: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:03:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:03:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:29 vm10.local ceph-mon[53712]: pgmap v1122: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:30.578 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:30.579 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:30.606 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:30.606 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:31 vm11.local ceph-mon[53973]: pgmap v1123: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:31 vm10.local ceph-mon[53712]: pgmap v1123: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:33.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:32 vm10.local ceph-mon[53712]: pgmap v1124: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:32 vm11.local ceph-mon[53973]: pgmap v1124: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:35.608 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:35.608 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:35.637 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:35.637 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:35 vm10.local ceph-mon[53712]: pgmap v1125: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:36.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:35 vm11.local ceph-mon[53973]: pgmap v1125: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:37 vm10.local ceph-mon[53712]: pgmap v1126: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:37 vm11.local ceph-mon[53973]: pgmap v1126: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:39.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:39 vm11.local ceph-mon[53973]: pgmap v1127: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:39 vm10.local ceph-mon[53712]: pgmap v1127: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:40.639 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:40.639 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:40.666 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:40.666 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:41.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:41 vm11.local ceph-mon[53973]: pgmap v1128: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:41 vm10.local ceph-mon[53712]: pgmap v1128: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:43 vm11.local ceph-mon[53973]: pgmap v1129: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:43.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:43 vm10.local ceph-mon[53712]: pgmap v1129: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:45.668 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:45.668 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:45.694 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:45.695 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:45.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:45 vm11.local ceph-mon[53973]: pgmap v1130: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:45 vm10.local ceph-mon[53712]: pgmap v1130: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:47 vm11.local ceph-mon[53973]: pgmap v1131: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:47 vm10.local ceph-mon[53712]: pgmap v1131: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:49 vm11.local ceph-mon[53973]: pgmap v1132: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:49.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:49 vm10.local ceph-mon[53712]: pgmap v1132: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:50.696 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:50.697 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:50.724 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:50.724 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:51 vm11.local ceph-mon[53973]: pgmap v1133: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:51 vm10.local ceph-mon[53712]: pgmap v1133: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:53 vm11.local ceph-mon[53973]: pgmap v1134: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:53 vm10.local ceph-mon[53712]: pgmap v1134: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:55.725 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:03:55.726 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:03:55.802 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:03:55.802 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:03:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:55 vm11.local ceph-mon[53973]: pgmap v1135: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:55 vm10.local ceph-mon[53712]: pgmap v1135: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:57 vm11.local ceph-mon[53973]: pgmap v1136: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:57 vm10.local ceph-mon[53712]: pgmap v1136: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:03:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:03:59 vm11.local ceph-mon[53973]: pgmap v1137: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:03:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:03:59 vm10.local ceph-mon[53712]: pgmap v1137: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:00.804 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:00.804 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:00.829 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:00.830 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:01 vm11.local ceph-mon[53973]: pgmap v1138: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:01 vm10.local ceph-mon[53712]: pgmap v1138: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:03 vm10.local ceph-mon[53712]: pgmap v1139: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:04.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:03 vm11.local ceph-mon[53973]: pgmap v1139: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:05.831 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:05.832 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:05 vm11.local ceph-mon[53973]: pgmap v1140: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:05.859 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:05.860 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:05 vm10.local ceph-mon[53712]: pgmap v1140: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:07 vm11.local ceph-mon[53973]: pgmap v1141: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:07 vm10.local ceph-mon[53712]: pgmap v1141: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:09 vm11.local ceph-mon[53973]: pgmap v1142: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:04:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:04:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:04:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:04:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:04:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:09 vm10.local ceph-mon[53712]: pgmap v1142: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:04:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:04:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:04:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:04:09.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:04:10.862 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:10.862 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:10.896 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:10.897 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:11 vm11.local ceph-mon[53973]: pgmap v1143: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:11.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:11 vm10.local ceph-mon[53712]: pgmap v1143: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:13 vm11.local ceph-mon[53973]: pgmap v1144: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:13 vm10.local ceph-mon[53712]: pgmap v1144: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:15 vm11.local ceph-mon[53973]: pgmap v1145: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:15.898 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:15.899 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:15.924 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:15.925 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:15 vm10.local ceph-mon[53712]: pgmap v1145: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:17 vm11.local ceph-mon[53973]: pgmap v1146: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:17 vm10.local ceph-mon[53712]: pgmap v1146: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:19 vm11.local ceph-mon[53973]: pgmap v1147: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:19 vm10.local ceph-mon[53712]: pgmap v1147: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:20.926 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:20.927 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:20.953 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:20.953 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:21 vm11.local ceph-mon[53973]: pgmap v1148: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:21 vm10.local ceph-mon[53712]: pgmap v1148: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:23 vm11.local ceph-mon[53973]: pgmap v1149: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:23 vm10.local ceph-mon[53712]: pgmap v1149: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:25 vm11.local ceph-mon[53973]: pgmap v1150: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:25 vm10.local ceph-mon[53712]: pgmap v1150: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:25.954 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:25.955 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:25.982 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:25.982 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:27 vm11.local ceph-mon[53973]: pgmap v1151: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:27 vm10.local ceph-mon[53712]: pgmap v1151: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:29.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:04:29.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:04:29.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:29 vm11.local ceph-mon[53973]: pgmap v1152: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:04:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:04:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:29 vm10.local ceph-mon[53712]: pgmap v1152: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:30.984 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:30.984 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:31.147 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:31.147 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:31.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:31 vm10.local ceph-mon[53712]: pgmap v1153: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:31 vm11.local ceph-mon[53973]: pgmap v1153: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:32 vm10.local ceph-mon[53712]: pgmap v1154: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:33.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:32 vm11.local ceph-mon[53973]: pgmap v1154: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:34.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:34 vm10.local ceph-mon[53712]: pgmap v1155: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:35.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:34 vm11.local ceph-mon[53973]: pgmap v1155: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:36.149 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:36.149 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:36.175 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:36.176 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:37 vm11.local ceph-mon[53973]: pgmap v1156: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:37 vm10.local ceph-mon[53712]: pgmap v1156: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:39 vm11.local ceph-mon[53973]: pgmap v1157: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:39 vm10.local ceph-mon[53712]: pgmap v1157: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:41.177 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:41.178 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:41.203 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:41.203 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:41 vm11.local ceph-mon[53973]: pgmap v1158: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:41 vm10.local ceph-mon[53712]: pgmap v1158: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:43 vm11.local ceph-mon[53973]: pgmap v1159: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:43 vm10.local ceph-mon[53712]: pgmap v1159: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:45 vm10.local ceph-mon[53712]: pgmap v1160: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:45 vm11.local ceph-mon[53973]: pgmap v1160: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:46.204 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:46.205 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:46.258 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:46.259 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:46.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:46 vm10.local ceph-mon[53712]: pgmap v1161: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:46 vm11.local ceph-mon[53973]: pgmap v1161: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:49 vm11.local ceph-mon[53973]: pgmap v1162: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:49.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:49 vm10.local ceph-mon[53712]: pgmap v1162: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:51.260 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:51.261 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:51.286 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:51.287 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:51 vm11.local ceph-mon[53973]: pgmap v1163: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:51.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:51 vm10.local ceph-mon[53712]: pgmap v1163: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:53 vm11.local ceph-mon[53973]: pgmap v1164: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:53 vm10.local ceph-mon[53712]: pgmap v1164: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:55 vm11.local ceph-mon[53973]: pgmap v1165: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:55 vm10.local ceph-mon[53712]: pgmap v1165: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:56.288 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:04:56.289 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:04:56.314 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:04:56.314 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:04:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:57 vm11.local ceph-mon[53973]: pgmap v1166: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:57 vm10.local ceph-mon[53712]: pgmap v1166: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:04:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:04:59 vm11.local ceph-mon[53973]: pgmap v1167: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:04:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:04:59 vm10.local ceph-mon[53712]: pgmap v1167: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:01.315 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:01.316 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:01.342 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:01.342 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:01 vm11.local ceph-mon[53973]: pgmap v1168: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:01 vm10.local ceph-mon[53712]: pgmap v1168: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:03 vm10.local ceph-mon[53712]: pgmap v1169: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:03 vm11.local ceph-mon[53973]: pgmap v1169: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:04.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:04 vm10.local ceph-mon[53712]: pgmap v1170: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:04 vm11.local ceph-mon[53973]: pgmap v1170: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:06.344 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:06.344 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:06.372 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:06.372 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:07 vm11.local ceph-mon[53973]: pgmap v1171: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:07 vm10.local ceph-mon[53712]: pgmap v1171: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:09 vm11.local ceph-mon[53973]: pgmap v1172: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:05:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:05:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:05:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:05:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:05:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:09 vm10.local ceph-mon[53712]: pgmap v1172: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:05:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:05:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:05:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:05:09.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:05:11.373 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:11.374 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:11.400 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:11.400 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:11 vm11.local ceph-mon[53973]: pgmap v1173: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:11.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:11 vm10.local ceph-mon[53712]: pgmap v1173: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:13 vm11.local ceph-mon[53973]: pgmap v1174: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:13 vm10.local ceph-mon[53712]: pgmap v1174: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:15 vm11.local ceph-mon[53973]: pgmap v1175: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:15 vm10.local ceph-mon[53712]: pgmap v1175: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:16.402 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:16.402 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:16.427 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:16.428 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:17 vm11.local ceph-mon[53973]: pgmap v1176: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:17 vm10.local ceph-mon[53712]: pgmap v1176: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:19 vm11.local ceph-mon[53973]: pgmap v1177: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:05:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:19 vm10.local ceph-mon[53712]: pgmap v1177: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:05:21.429 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:21.430 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:21.455 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:21.456 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:21 vm11.local ceph-mon[53973]: pgmap v1178: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:05:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:21 vm10.local ceph-mon[53712]: pgmap v1178: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:05:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:23 vm11.local ceph-mon[53973]: pgmap v1179: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:23 vm10.local ceph-mon[53712]: pgmap v1179: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:25 vm11.local ceph-mon[53973]: pgmap v1180: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:05:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:25 vm10.local ceph-mon[53712]: pgmap v1180: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:05:26.457 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:26.458 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:26.484 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:26.485 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:27 vm11.local ceph-mon[53973]: pgmap v1181: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:27 vm10.local ceph-mon[53712]: pgmap v1181: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:05:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:05:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:29 vm11.local ceph-mon[53973]: pgmap v1182: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:05:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:05:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:29 vm10.local ceph-mon[53712]: pgmap v1182: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:31.486 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:31.486 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:31.513 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:31.514 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:31 vm11.local ceph-mon[53973]: pgmap v1183: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:31 vm10.local ceph-mon[53712]: pgmap v1183: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:33 vm11.local ceph-mon[53973]: pgmap v1184: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:33 vm10.local ceph-mon[53712]: pgmap v1184: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:35 vm11.local ceph-mon[53973]: pgmap v1185: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:35 vm10.local ceph-mon[53712]: pgmap v1185: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:36.515 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:36.516 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:36.542 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:36.543 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:37 vm11.local ceph-mon[53973]: pgmap v1186: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:37 vm10.local ceph-mon[53712]: pgmap v1186: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:39 vm11.local ceph-mon[53973]: pgmap v1187: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:39.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:39 vm10.local ceph-mon[53712]: pgmap v1187: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:41.544 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:41.545 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:41.573 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:41.574 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:41 vm11.local ceph-mon[53973]: pgmap v1188: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:41 vm10.local ceph-mon[53712]: pgmap v1188: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:43 vm11.local ceph-mon[53973]: pgmap v1189: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:43 vm10.local ceph-mon[53712]: pgmap v1189: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:45 vm11.local ceph-mon[53973]: pgmap v1190: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:45 vm10.local ceph-mon[53712]: pgmap v1190: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:46.575 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:46.576 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:46.602 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:46.602 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:47 vm11.local ceph-mon[53973]: pgmap v1191: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:47 vm10.local ceph-mon[53712]: pgmap v1191: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:49 vm11.local ceph-mon[53973]: pgmap v1192: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:49.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:49 vm10.local ceph-mon[53712]: pgmap v1192: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:51.603 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:51.604 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:51.629 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:51.630 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:51 vm11.local ceph-mon[53973]: pgmap v1193: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:51 vm10.local ceph-mon[53712]: pgmap v1193: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:53 vm11.local ceph-mon[53973]: pgmap v1194: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:53 vm10.local ceph-mon[53712]: pgmap v1194: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:55 vm11.local ceph-mon[53973]: pgmap v1195: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:55 vm10.local ceph-mon[53712]: pgmap v1195: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:56.631 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:05:56.631 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:05:56.656 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:05:56.657 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:05:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:57 vm11.local ceph-mon[53973]: pgmap v1196: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:57 vm10.local ceph-mon[53712]: pgmap v1196: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:05:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:05:59 vm11.local ceph-mon[53973]: pgmap v1197: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:05:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:05:59 vm10.local ceph-mon[53712]: pgmap v1197: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:01.659 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:01.659 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:01.686 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:01.687 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:01 vm11.local ceph-mon[53973]: pgmap v1198: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:01.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:01 vm10.local ceph-mon[53712]: pgmap v1198: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:03 vm11.local ceph-mon[53973]: pgmap v1199: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:03.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:03 vm10.local ceph-mon[53712]: pgmap v1199: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:05 vm11.local ceph-mon[53973]: pgmap v1200: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:05 vm10.local ceph-mon[53712]: pgmap v1200: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:06.688 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:06.689 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:06.715 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:06.715 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:07 vm11.local ceph-mon[53973]: pgmap v1201: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:07 vm10.local ceph-mon[53712]: pgmap v1201: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:09.288 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:09 vm11.local ceph-mon[53973]: pgmap v1202: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:09.288 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:06:09.288 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:06:09.288 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:06:09.288 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:06:09.288 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:06:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:09 vm10.local ceph-mon[53712]: pgmap v1202: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:06:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:06:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:06:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:06:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:06:11.716 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:11.717 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:11.742 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:11.743 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:11 vm11.local ceph-mon[53973]: pgmap v1203: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:11 vm10.local ceph-mon[53712]: pgmap v1203: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:13 vm11.local ceph-mon[53973]: pgmap v1204: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:13 vm10.local ceph-mon[53712]: pgmap v1204: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:15 vm11.local ceph-mon[53973]: pgmap v1205: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:15 vm10.local ceph-mon[53712]: pgmap v1205: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:16.744 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:16.744 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:16.772 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:16.772 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:17 vm11.local ceph-mon[53973]: pgmap v1206: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:17 vm10.local ceph-mon[53712]: pgmap v1206: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:19 vm11.local ceph-mon[53973]: pgmap v1207: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:19 vm10.local ceph-mon[53712]: pgmap v1207: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:21.774 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:21.775 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:21.801 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:21.801 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:21 vm11.local ceph-mon[53973]: pgmap v1208: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:21 vm10.local ceph-mon[53712]: pgmap v1208: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:23 vm11.local ceph-mon[53973]: pgmap v1209: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:23 vm10.local ceph-mon[53712]: pgmap v1209: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:25 vm11.local ceph-mon[53973]: pgmap v1210: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:25 vm10.local ceph-mon[53712]: pgmap v1210: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:26.803 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:26.804 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:26.833 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:26.833 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:27 vm11.local ceph-mon[53973]: pgmap v1211: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:27 vm10.local ceph-mon[53712]: pgmap v1211: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:06:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:06:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:29 vm11.local ceph-mon[53973]: pgmap v1212: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:06:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:06:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:29 vm10.local ceph-mon[53712]: pgmap v1212: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:31.835 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:31.836 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:31 vm11.local ceph-mon[53973]: pgmap v1213: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:31.862 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:31.863 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:31 vm10.local ceph-mon[53712]: pgmap v1213: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:33 vm11.local ceph-mon[53973]: pgmap v1214: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:33.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:33 vm10.local ceph-mon[53712]: pgmap v1214: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:35 vm11.local ceph-mon[53973]: pgmap v1215: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:35 vm10.local ceph-mon[53712]: pgmap v1215: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:36.864 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:36.865 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:36.890 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:36.890 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:37 vm11.local ceph-mon[53973]: pgmap v1216: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:37 vm10.local ceph-mon[53712]: pgmap v1216: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:39 vm11.local ceph-mon[53973]: pgmap v1217: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:39.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:39 vm10.local ceph-mon[53712]: pgmap v1217: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:41 vm11.local ceph-mon[53973]: pgmap v1218: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:41.892 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:41.900 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:41 vm10.local ceph-mon[53712]: pgmap v1218: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:41.936 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:41.937 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:43 vm11.local ceph-mon[53973]: pgmap v1219: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:43 vm10.local ceph-mon[53712]: pgmap v1219: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:45 vm11.local ceph-mon[53973]: pgmap v1220: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:45 vm10.local ceph-mon[53712]: pgmap v1220: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:46.938 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:46.939 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:46.965 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:46.966 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:47 vm10.local ceph-mon[53712]: pgmap v1221: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:47 vm11.local ceph-mon[53973]: pgmap v1221: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:49.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:49 vm10.local ceph-mon[53712]: pgmap v1222: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:49.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:49 vm11.local ceph-mon[53973]: pgmap v1222: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:51 vm11.local ceph-mon[53973]: pgmap v1223: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:51 vm10.local ceph-mon[53712]: pgmap v1223: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:51.967 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:51.968 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:51.994 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:51.995 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:53 vm11.local ceph-mon[53973]: pgmap v1224: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:53 vm10.local ceph-mon[53712]: pgmap v1224: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:55 vm11.local ceph-mon[53973]: pgmap v1225: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:55.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:55 vm10.local ceph-mon[53712]: pgmap v1225: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:56.996 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:06:56.997 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:06:57.025 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:06:57.025 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:06:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:57 vm11.local ceph-mon[53973]: pgmap v1226: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:57 vm10.local ceph-mon[53712]: pgmap v1226: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:06:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:06:59 vm11.local ceph-mon[53973]: pgmap v1227: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:06:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:06:59 vm10.local ceph-mon[53712]: pgmap v1227: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:01 vm11.local ceph-mon[53973]: pgmap v1228: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:01 vm10.local ceph-mon[53712]: pgmap v1228: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:02.027 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:02.027 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:02.057 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:02.057 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:03 vm11.local ceph-mon[53973]: pgmap v1229: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:03 vm10.local ceph-mon[53712]: pgmap v1229: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:05 vm10.local ceph-mon[53712]: pgmap v1230: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:06.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:05 vm11.local ceph-mon[53973]: pgmap v1230: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:06.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:06 vm10.local ceph-mon[53712]: pgmap v1231: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:07.059 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:07.059 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:07.086 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:07.086 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:06 vm11.local ceph-mon[53973]: pgmap v1231: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:09 vm11.local ceph-mon[53973]: pgmap v1232: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:07:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:07:09.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:09 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:07:09.638 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:09 vm10.local ceph-mon[53712]: pgmap v1232: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:09.638 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:07:09.638 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:07:09.638 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:09 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:07:10.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:07:10.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:07:10.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:07:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:07:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:07:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:07:11.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:11 vm10.local ceph-mon[53712]: pgmap v1233: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:11 vm11.local ceph-mon[53973]: pgmap v1233: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:12.087 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:12.116 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:12.182 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:12.182 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:13 vm11.local ceph-mon[53973]: pgmap v1234: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:13 vm10.local ceph-mon[53712]: pgmap v1234: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:14 vm11.local ceph-mon[53973]: pgmap v1235: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:14 vm10.local ceph-mon[53712]: pgmap v1235: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:17.184 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:17.184 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:17.227 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:17.228 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:17 vm11.local ceph-mon[53973]: pgmap v1236: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:17 vm10.local ceph-mon[53712]: pgmap v1236: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:19.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:19 vm10.local ceph-mon[53712]: pgmap v1237: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:19.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:19 vm11.local ceph-mon[53973]: pgmap v1237: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:21 vm10.local ceph-mon[53712]: pgmap v1238: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:21 vm11.local ceph-mon[53973]: pgmap v1238: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:22.229 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:22.230 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:22.402 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:22.403 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:22 vm10.local ceph-mon[53712]: pgmap v1239: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:22 vm11.local ceph-mon[53973]: pgmap v1239: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:25.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:25 vm11.local ceph-mon[53973]: pgmap v1240: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:25.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:25 vm10.local ceph-mon[53712]: pgmap v1240: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:27.404 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:27.405 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:27.433 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:27.433 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:27 vm11.local ceph-mon[53973]: pgmap v1241: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:27 vm10.local ceph-mon[53712]: pgmap v1241: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:28.827 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:07:28.827 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:07:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:07:28.858 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:07:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:29 vm11.local ceph-mon[53973]: pgmap v1242: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:29 vm10.local ceph-mon[53712]: pgmap v1242: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:31 vm10.local ceph-mon[53712]: pgmap v1243: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:31 vm11.local ceph-mon[53973]: pgmap v1243: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:32.434 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:32.435 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:32.461 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:32.462 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:33.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:33 vm10.local ceph-mon[53712]: pgmap v1244: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:33 vm11.local ceph-mon[53973]: pgmap v1244: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:35 vm10.local ceph-mon[53712]: pgmap v1245: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:36.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:35 vm11.local ceph-mon[53973]: pgmap v1245: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:37.464 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:37.464 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:37.490 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:37.491 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:37.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:37 vm10.local ceph-mon[53712]: pgmap v1246: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:37 vm11.local ceph-mon[53973]: pgmap v1246: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:39.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:39 vm10.local ceph-mon[53712]: pgmap v1247: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:39 vm11.local ceph-mon[53973]: pgmap v1247: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:41 vm11.local ceph-mon[53973]: pgmap v1248: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:41 vm10.local ceph-mon[53712]: pgmap v1248: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:42.493 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:42.493 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:42.521 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:42.521 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:43 vm11.local ceph-mon[53973]: pgmap v1249: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:43 vm10.local ceph-mon[53712]: pgmap v1249: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:45.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:45 vm11.local ceph-mon[53973]: pgmap v1250: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:45 vm10.local ceph-mon[53712]: pgmap v1250: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:47.523 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:47.523 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:47.550 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:47.550 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:47 vm11.local ceph-mon[53973]: pgmap v1251: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:47 vm10.local ceph-mon[53712]: pgmap v1251: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:49.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:49 vm11.local ceph-mon[53973]: pgmap v1252: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:49.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:49 vm10.local ceph-mon[53712]: pgmap v1252: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:51 vm11.local ceph-mon[53973]: pgmap v1253: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:51 vm10.local ceph-mon[53712]: pgmap v1253: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:52.552 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:52.552 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:52.580 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:52.580 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:53.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:53 vm11.local ceph-mon[53973]: pgmap v1254: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:53 vm10.local ceph-mon[53712]: pgmap v1254: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:55 vm11.local ceph-mon[53973]: pgmap v1255: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:55 vm10.local ceph-mon[53712]: pgmap v1255: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:57.582 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:07:57.582 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:07:57.609 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:07:57.609 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:07:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:57 vm11.local ceph-mon[53973]: pgmap v1256: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:57 vm10.local ceph-mon[53712]: pgmap v1256: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:07:59.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:07:59 vm11.local ceph-mon[53973]: pgmap v1257: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:07:59.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:07:59 vm10.local ceph-mon[53712]: pgmap v1257: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:01 vm11.local ceph-mon[53973]: pgmap v1258: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:01 vm10.local ceph-mon[53712]: pgmap v1258: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:02.611 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:02.611 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:02.641 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:02.641 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:03 vm11.local ceph-mon[53973]: pgmap v1259: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:03 vm10.local ceph-mon[53712]: pgmap v1259: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:05 vm11.local ceph-mon[53973]: pgmap v1260: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:05 vm10.local ceph-mon[53712]: pgmap v1260: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:07.642 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:07.643 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:07.669 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:07.669 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:07.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:07 vm11.local ceph-mon[53973]: pgmap v1261: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:07 vm10.local ceph-mon[53712]: pgmap v1261: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:09.704 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:09 vm11.local ceph-mon[53973]: pgmap v1262: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:09.708 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:09 vm10.local ceph-mon[53712]: pgmap v1262: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:08:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:08:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:08:10.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:08:10.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:08:10.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:08:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:08:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:08:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:08:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:08:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:08:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:08:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:11 vm11.local ceph-mon[53973]: pgmap v1263: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:11.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:11 vm10.local ceph-mon[53712]: pgmap v1263: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:12.671 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:12.671 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:12.697 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:12.697 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:13.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:13 vm11.local ceph-mon[53973]: pgmap v1264: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:13 vm10.local ceph-mon[53712]: pgmap v1264: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:15 vm11.local ceph-mon[53973]: pgmap v1265: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:15 vm10.local ceph-mon[53712]: pgmap v1265: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:17.698 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:17.699 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:17.725 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:17.726 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:17 vm11.local ceph-mon[53973]: pgmap v1266: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:17 vm10.local ceph-mon[53712]: pgmap v1266: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:19.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:19 vm11.local ceph-mon[53973]: pgmap v1267: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:19.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:19 vm10.local ceph-mon[53712]: pgmap v1267: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:21 vm11.local ceph-mon[53973]: pgmap v1268: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:21 vm10.local ceph-mon[53712]: pgmap v1268: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:22.727 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:22.728 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:22.753 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:22.754 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:23 vm10.local ceph-mon[53712]: pgmap v1269: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:23 vm11.local ceph-mon[53973]: pgmap v1269: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:24 vm11.local ceph-mon[53973]: pgmap v1270: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:24 vm10.local ceph-mon[53712]: pgmap v1270: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:27.755 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:27.756 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:27.783 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:27.783 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:27 vm11.local ceph-mon[53973]: pgmap v1271: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:27 vm10.local ceph-mon[53712]: pgmap v1271: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:28.834 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:08:28.834 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:08:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:08:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:08:29.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:29 vm11.local ceph-mon[53973]: pgmap v1272: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:29 vm10.local ceph-mon[53712]: pgmap v1272: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:31 vm11.local ceph-mon[53973]: pgmap v1273: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:31.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:31 vm10.local ceph-mon[53712]: pgmap v1273: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:32.784 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:32.785 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:33.074 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:33.075 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:33 vm11.local ceph-mon[53973]: pgmap v1274: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:33 vm10.local ceph-mon[53712]: pgmap v1274: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:35 vm11.local ceph-mon[53973]: pgmap v1275: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:35 vm10.local ceph-mon[53712]: pgmap v1275: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:37 vm10.local ceph-mon[53712]: pgmap v1276: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:38.076 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:38.077 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:37 vm11.local ceph-mon[53973]: pgmap v1276: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:38.105 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:38.105 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:39 vm11.local ceph-mon[53973]: pgmap v1277: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:39 vm10.local ceph-mon[53712]: pgmap v1277: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:41 vm11.local ceph-mon[53973]: pgmap v1278: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:41 vm10.local ceph-mon[53712]: pgmap v1278: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:43.107 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:43.107 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:43.145 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:43.145 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:43 vm11.local ceph-mon[53973]: pgmap v1279: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:43 vm10.local ceph-mon[53712]: pgmap v1279: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:45 vm11.local ceph-mon[53973]: pgmap v1280: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:45.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:45 vm10.local ceph-mon[53712]: pgmap v1280: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:47.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:47 vm11.local ceph-mon[53973]: pgmap v1281: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:47 vm10.local ceph-mon[53712]: pgmap v1281: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:48.146 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:48.147 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:48.173 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:48.174 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:49.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:49 vm11.local ceph-mon[53973]: pgmap v1282: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:49.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:49 vm10.local ceph-mon[53712]: pgmap v1282: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:51 vm11.local ceph-mon[53973]: pgmap v1283: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:51 vm10.local ceph-mon[53712]: pgmap v1283: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:53.175 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:53.176 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:53.283 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:53.284 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:53 vm11.local ceph-mon[53973]: pgmap v1284: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:53 vm10.local ceph-mon[53712]: pgmap v1284: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:54 vm11.local ceph-mon[53973]: pgmap v1285: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:54 vm10.local ceph-mon[53712]: pgmap v1285: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:57.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:57 vm11.local ceph-mon[53973]: pgmap v1286: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:57 vm10.local ceph-mon[53712]: pgmap v1286: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:08:58.285 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:08:58.285 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:08:58.311 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:08:58.312 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:08:59.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:08:59 vm11.local ceph-mon[53973]: pgmap v1287: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:08:59.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:08:59 vm10.local ceph-mon[53712]: pgmap v1287: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:01.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:01 vm11.local ceph-mon[53973]: pgmap v1288: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:01.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:01 vm10.local ceph-mon[53712]: pgmap v1288: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:03.313 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:03.313 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:03.346 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:03.346 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:03.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:03 vm11.local ceph-mon[53973]: pgmap v1289: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:03.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:03 vm10.local ceph-mon[53712]: pgmap v1289: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:05.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:05 vm11.local ceph-mon[53973]: pgmap v1290: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:05 vm10.local ceph-mon[53712]: pgmap v1290: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:07.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:07 vm10.local ceph-mon[53712]: pgmap v1291: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:07 vm11.local ceph-mon[53973]: pgmap v1291: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:08.348 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:08.348 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:08.438 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:08.439 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:09 vm11.local ceph-mon[53973]: pgmap v1292: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:09 vm10.local ceph-mon[53712]: pgmap v1292: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:10.506 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:09:10.506 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:09:10.506 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:09:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:09:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:09:10.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:09:10.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:09:10.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:09:10.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:09:10.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:09:10.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:09:10.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:09:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:11 vm11.local ceph-mon[53973]: pgmap v1293: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:11 vm10.local ceph-mon[53712]: pgmap v1293: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:13.440 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:13.441 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:13.467 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:13.467 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:13 vm11.local ceph-mon[53973]: pgmap v1294: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:13 vm10.local ceph-mon[53712]: pgmap v1294: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:14 vm10.local ceph-mon[53712]: pgmap v1295: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:14 vm11.local ceph-mon[53973]: pgmap v1295: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:17 vm11.local ceph-mon[53973]: pgmap v1296: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:17 vm10.local ceph-mon[53712]: pgmap v1296: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:18.469 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:18.469 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:18.494 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:18.495 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:19.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:19 vm11.local ceph-mon[53973]: pgmap v1297: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:19.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:19 vm10.local ceph-mon[53712]: pgmap v1297: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:22.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:21 vm11.local ceph-mon[53973]: pgmap v1298: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:22.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:21 vm10.local ceph-mon[53712]: pgmap v1298: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:22 vm11.local ceph-mon[53973]: pgmap v1299: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:22 vm10.local ceph-mon[53712]: pgmap v1299: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:23.496 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:23.497 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:23.522 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:23.523 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:25 vm11.local ceph-mon[53973]: pgmap v1300: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:25 vm10.local ceph-mon[53712]: pgmap v1300: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:27.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:27 vm11.local ceph-mon[53973]: pgmap v1301: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:27.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:27 vm10.local ceph-mon[53712]: pgmap v1301: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:28.524 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:28.524 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:28.550 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:28.550 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:28.838 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:09:28.838 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:09:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:09:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:09:29.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:29 vm11.local ceph-mon[53973]: pgmap v1302: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:29 vm10.local ceph-mon[53712]: pgmap v1302: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:31.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:31 vm11.local ceph-mon[53973]: pgmap v1303: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:31 vm10.local ceph-mon[53712]: pgmap v1303: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:33.552 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:33.552 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:33.581 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:33.581 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:33 vm11.local ceph-mon[53973]: pgmap v1304: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:33 vm10.local ceph-mon[53712]: pgmap v1304: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:35 vm11.local ceph-mon[53973]: pgmap v1305: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:35 vm10.local ceph-mon[53712]: pgmap v1305: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:37.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:37 vm11.local ceph-mon[53973]: pgmap v1306: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:37 vm10.local ceph-mon[53712]: pgmap v1306: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:38.582 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:38.583 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:38.611 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:38.611 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:39 vm11.local ceph-mon[53973]: pgmap v1307: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:39 vm10.local ceph-mon[53712]: pgmap v1307: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:41 vm10.local ceph-mon[53712]: pgmap v1308: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:41 vm11.local ceph-mon[53973]: pgmap v1308: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:43.612 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:43.613 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:43.639 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:43.639 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:43 vm10.local ceph-mon[53712]: pgmap v1309: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:43 vm11.local ceph-mon[53973]: pgmap v1309: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:45 vm10.local ceph-mon[53712]: pgmap v1310: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:45 vm11.local ceph-mon[53973]: pgmap v1310: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:47 vm10.local ceph-mon[53712]: pgmap v1311: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:47 vm11.local ceph-mon[53973]: pgmap v1311: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:48.640 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:48.641 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:48.667 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:48.667 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:49.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:49 vm10.local ceph-mon[53712]: pgmap v1312: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:49 vm11.local ceph-mon[53973]: pgmap v1312: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:51 vm10.local ceph-mon[53712]: pgmap v1313: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:51 vm11.local ceph-mon[53973]: pgmap v1313: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:53.669 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:53.669 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:53.695 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:53.695 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:53 vm10.local ceph-mon[53712]: pgmap v1314: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:53 vm11.local ceph-mon[53973]: pgmap v1314: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:55.892 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:55 vm11.local ceph-mon[53973]: pgmap v1315: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:55 vm10.local ceph-mon[53712]: pgmap v1315: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:09:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:57 vm10.local ceph-mon[53712]: pgmap v1316: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:57 vm11.local ceph-mon[53973]: pgmap v1316: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:09:58.697 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:09:58.697 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:09:58.724 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:09:58.725 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:09:59.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:09:59 vm10.local ceph-mon[53712]: pgmap v1317: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:00.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:09:59 vm11.local ceph-mon[53973]: pgmap v1317: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:00.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:00 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T18:10:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:00 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T18:10:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:01 vm10.local ceph-mon[53712]: pgmap v1318: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:01 vm11.local ceph-mon[53973]: pgmap v1318: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:03.726 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:03.727 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:03.754 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:03.755 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:03 vm10.local ceph-mon[53712]: pgmap v1319: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:03 vm11.local ceph-mon[53973]: pgmap v1319: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:06.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:05 vm11.local ceph-mon[53973]: pgmap v1320: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:06.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:05 vm10.local ceph-mon[53712]: pgmap v1320: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:06 vm10.local ceph-mon[53712]: pgmap v1321: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:06 vm11.local ceph-mon[53973]: pgmap v1321: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:08.756 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:08.757 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:08.782 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:08.783 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:09.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:09 vm11.local ceph-mon[53973]: pgmap v1322: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:09 vm10.local ceph-mon[53712]: pgmap v1322: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:10.683 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:10:10.683 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:10:10.683 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:10 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:10:10.709 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:10:10.709 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:10:10.709 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:10 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:10:11.740 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:11 vm10.local ceph-mon[53712]: pgmap v1323: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:11.833 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:11 vm11.local ceph-mon[53973]: pgmap v1323: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:12.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:10:12.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:10:12.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:10:12.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:10:12.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:10:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:10:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:10:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:10:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:10:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:10:13.784 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:13.785 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:13.809 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:13.810 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:13 vm10.local ceph-mon[53712]: pgmap v1324: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:13 vm11.local ceph-mon[53973]: pgmap v1324: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:15 vm10.local ceph-mon[53712]: pgmap v1325: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:16.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:15 vm11.local ceph-mon[53973]: pgmap v1325: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:17.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:16 vm11.local ceph-mon[53973]: pgmap v1326: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:16 vm10.local ceph-mon[53712]: pgmap v1326: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:18.811 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:18.812 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:18.846 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:18.847 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:19.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:19 vm10.local ceph-mon[53712]: pgmap v1327: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:19 vm11.local ceph-mon[53973]: pgmap v1327: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:21.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:21 vm10.local ceph-mon[53712]: pgmap v1328: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:21 vm11.local ceph-mon[53973]: pgmap v1328: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:23.848 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:23.849 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:23.875 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:23.876 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:23 vm10.local ceph-mon[53712]: pgmap v1329: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:23 vm11.local ceph-mon[53973]: pgmap v1329: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:25.892 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:25 vm11.local ceph-mon[53973]: pgmap v1330: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:25.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:25 vm10.local ceph-mon[53712]: pgmap v1330: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:27.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:27 vm10.local ceph-mon[53712]: pgmap v1331: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:28.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:27 vm11.local ceph-mon[53973]: pgmap v1331: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:28.878 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:28.878 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:28.906 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:28.906 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:10:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:10:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:10:29.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:10:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:29 vm10.local ceph-mon[53712]: pgmap v1332: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:29 vm11.local ceph-mon[53973]: pgmap v1332: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:31.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:31 vm10.local ceph-mon[53712]: pgmap v1333: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:31 vm11.local ceph-mon[53973]: pgmap v1333: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:33.908 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:33.908 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:33.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:33 vm10.local ceph-mon[53712]: pgmap v1334: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:33.936 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:33.937 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:33 vm11.local ceph-mon[53973]: pgmap v1334: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:36.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:35 vm11.local ceph-mon[53973]: pgmap v1335: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:36.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:35 vm10.local ceph-mon[53712]: pgmap v1335: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:36.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:36 vm11.local ceph-mon[53973]: pgmap v1336: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:37.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:36 vm10.local ceph-mon[53712]: pgmap v1336: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:38.938 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:38.938 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:38.964 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:38.965 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:39.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:39 vm11.local ceph-mon[53973]: pgmap v1337: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:39 vm10.local ceph-mon[53712]: pgmap v1337: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:41 vm10.local ceph-mon[53712]: pgmap v1338: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:41 vm11.local ceph-mon[53973]: pgmap v1338: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:43 vm10.local ceph-mon[53712]: pgmap v1339: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:43.966 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:43.966 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:43.993 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:43.994 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:43 vm11.local ceph-mon[53973]: pgmap v1339: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:45 vm10.local ceph-mon[53712]: pgmap v1340: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:45 vm11.local ceph-mon[53973]: pgmap v1340: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:46 vm11.local ceph-mon[53973]: pgmap v1341: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:46 vm10.local ceph-mon[53712]: pgmap v1341: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:48.995 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:49.003 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:49.074 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:49.075 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:48 vm11.local ceph-mon[53973]: pgmap v1342: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:48 vm10.local ceph-mon[53712]: pgmap v1342: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:51.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:51 vm11.local ceph-mon[53973]: pgmap v1343: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:51 vm10.local ceph-mon[53712]: pgmap v1343: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:53 vm10.local ceph-mon[53712]: pgmap v1344: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:54.076 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:54.077 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:53 vm11.local ceph-mon[53973]: pgmap v1344: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:54.103 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:54.104 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:55.893 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:55 vm11.local ceph-mon[53973]: pgmap v1345: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:55 vm10.local ceph-mon[53712]: pgmap v1345: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:10:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:57 vm10.local ceph-mon[53712]: pgmap v1346: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:57 vm11.local ceph-mon[53973]: pgmap v1346: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:10:59.105 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:10:59.106 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:10:59.131 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:10:59.131 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:10:59.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:10:59 vm10.local ceph-mon[53712]: pgmap v1347: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:10:59 vm11.local ceph-mon[53973]: pgmap v1347: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:01 vm10.local ceph-mon[53712]: pgmap v1348: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:01 vm11.local ceph-mon[53973]: pgmap v1348: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:03 vm10.local ceph-mon[53712]: pgmap v1349: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:03 vm11.local ceph-mon[53973]: pgmap v1349: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:04.133 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:04.133 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:04.161 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:04.162 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:05.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:05 vm10.local ceph-mon[53712]: pgmap v1350: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:06.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:05 vm11.local ceph-mon[53973]: pgmap v1350: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:07.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:07 vm10.local ceph-mon[53712]: pgmap v1351: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:07 vm11.local ceph-mon[53973]: pgmap v1351: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:09.163 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:09.163 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:09.190 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:09.190 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:09 vm10.local ceph-mon[53712]: pgmap v1352: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:09 vm11.local ceph-mon[53973]: pgmap v1352: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:11 vm10.local ceph-mon[53712]: pgmap v1353: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:11 vm11.local ceph-mon[53973]: pgmap v1353: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:12.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:11:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:11:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:11:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:11:12.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:12 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:11:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:11:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:11:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:11:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:11:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:12 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:11:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:13 vm10.local ceph-mon[53712]: pgmap v1354: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:13 vm11.local ceph-mon[53973]: pgmap v1354: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:14.191 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:14.192 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:14.219 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:14.219 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:14 vm11.local ceph-mon[53973]: pgmap v1355: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:14 vm10.local ceph-mon[53712]: pgmap v1355: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:17.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:17 vm11.local ceph-mon[53973]: pgmap v1356: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:17 vm10.local ceph-mon[53712]: pgmap v1356: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:19.220 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:19.221 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:19.247 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:19.248 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:19.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:19 vm10.local ceph-mon[53712]: pgmap v1357: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:19 vm11.local ceph-mon[53973]: pgmap v1357: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:21 vm11.local ceph-mon[53973]: pgmap v1358: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:21 vm10.local ceph-mon[53712]: pgmap v1358: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:22 vm11.local ceph-mon[53973]: pgmap v1359: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:22 vm10.local ceph-mon[53712]: pgmap v1359: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:24.249 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:24.249 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:24.275 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:24.275 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:25 vm10.local ceph-mon[53712]: pgmap v1360: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:26.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:25 vm11.local ceph-mon[53973]: pgmap v1360: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:27.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:27 vm10.local ceph-mon[53712]: pgmap v1361: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:27 vm11.local ceph-mon[53973]: pgmap v1361: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:11:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:11:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:11:29.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:11:29.277 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:29.277 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:29.307 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:29.307 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:29 vm10.local ceph-mon[53712]: pgmap v1362: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:30.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:29 vm11.local ceph-mon[53973]: pgmap v1362: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:31 vm10.local ceph-mon[53712]: pgmap v1363: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:31 vm11.local ceph-mon[53973]: pgmap v1363: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:33 vm10.local ceph-mon[53712]: pgmap v1364: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:33 vm11.local ceph-mon[53973]: pgmap v1364: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:34.309 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:34.309 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:34.336 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:34.336 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:35 vm10.local ceph-mon[53712]: pgmap v1365: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:36.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:35 vm11.local ceph-mon[53973]: pgmap v1365: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:37 vm10.local ceph-mon[53712]: pgmap v1366: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:37 vm11.local ceph-mon[53973]: pgmap v1366: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:39.338 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:39.338 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:39.368 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:39.369 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:39 vm10.local ceph-mon[53712]: pgmap v1367: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:39 vm11.local ceph-mon[53973]: pgmap v1367: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:40 vm11.local ceph-mon[53973]: pgmap v1368: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:40 vm10.local ceph-mon[53712]: pgmap v1368: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:43 vm10.local ceph-mon[53712]: pgmap v1369: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:43 vm11.local ceph-mon[53973]: pgmap v1369: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:44.371 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:44.371 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:44.397 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:44.398 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:45 vm10.local ceph-mon[53712]: pgmap v1370: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:45 vm11.local ceph-mon[53973]: pgmap v1370: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:47 vm10.local ceph-mon[53712]: pgmap v1371: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:47 vm11.local ceph-mon[53973]: pgmap v1371: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:49.399 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:49.399 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:49.426 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:49.427 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:49.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:49 vm10.local ceph-mon[53712]: pgmap v1372: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:49 vm11.local ceph-mon[53973]: pgmap v1372: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:51 vm10.local ceph-mon[53712]: pgmap v1373: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:51 vm11.local ceph-mon[53973]: pgmap v1373: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:53 vm10.local ceph-mon[53712]: pgmap v1374: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:53 vm11.local ceph-mon[53973]: pgmap v1374: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:54.428 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:54.429 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:54.456 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:54.456 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:55.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:55 vm10.local ceph-mon[53712]: pgmap v1375: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:56.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:55 vm11.local ceph-mon[53973]: pgmap v1375: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:11:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:57 vm10.local ceph-mon[53712]: pgmap v1376: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:57 vm11.local ceph-mon[53973]: pgmap v1376: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:11:59.457 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:11:59.458 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:11:59.484 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:11:59.484 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:11:59.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:11:59 vm10.local ceph-mon[53712]: pgmap v1377: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:11:59 vm11.local ceph-mon[53973]: pgmap v1377: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:02.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:01 vm11.local ceph-mon[53973]: pgmap v1378: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:01 vm10.local ceph-mon[53712]: pgmap v1378: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:02 vm11.local ceph-mon[53973]: pgmap v1379: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:03.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:02 vm10.local ceph-mon[53712]: pgmap v1379: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:04.486 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:04.486 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:04.514 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:04.515 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:05.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:05 vm10.local ceph-mon[53712]: pgmap v1380: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:06.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:05 vm11.local ceph-mon[53973]: pgmap v1380: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:07 vm10.local ceph-mon[53712]: pgmap v1381: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:07 vm11.local ceph-mon[53973]: pgmap v1381: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:09.516 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:09.516 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:09.543 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:09.543 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:09 vm10.local ceph-mon[53712]: pgmap v1382: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:10.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:09 vm11.local ceph-mon[53973]: pgmap v1382: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:11 vm10.local ceph-mon[53712]: pgmap v1383: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:11 vm11.local ceph-mon[53973]: pgmap v1383: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:13 vm10.local ceph-mon[53712]: pgmap v1384: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:12:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:12:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:12:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:12:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:12:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:13 vm11.local ceph-mon[53973]: pgmap v1384: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:12:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:12:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:12:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:12:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:12:14.545 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:14.545 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:14.571 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:14.571 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:15 vm10.local ceph-mon[53712]: pgmap v1385: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:16.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:15 vm11.local ceph-mon[53973]: pgmap v1385: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:17.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:17 vm10.local ceph-mon[53712]: pgmap v1386: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:17 vm11.local ceph-mon[53973]: pgmap v1386: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:19.573 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:19.573 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:19.600 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:19.600 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:19.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:19 vm10.local ceph-mon[53712]: pgmap v1387: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:19 vm11.local ceph-mon[53973]: pgmap v1387: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:20 vm11.local ceph-mon[53973]: pgmap v1388: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:20 vm10.local ceph-mon[53712]: pgmap v1388: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:23 vm10.local ceph-mon[53712]: pgmap v1389: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:23 vm11.local ceph-mon[53973]: pgmap v1389: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:24.601 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:24.602 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:24.629 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:24.630 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:25 vm10.local ceph-mon[53712]: pgmap v1390: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:26.038 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:25 vm11.local ceph-mon[53973]: pgmap v1390: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:26 vm11.local ceph-mon[53973]: pgmap v1391: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:26 vm10.local ceph-mon[53712]: pgmap v1391: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:28.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:12:28.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:12:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:12:28.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:12:29.631 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:29.632 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:29.658 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:29.659 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:29 vm10.local ceph-mon[53712]: pgmap v1392: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:29 vm11.local ceph-mon[53973]: pgmap v1392: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:31 vm10.local ceph-mon[53712]: pgmap v1393: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:31 vm11.local ceph-mon[53973]: pgmap v1393: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:33 vm11.local ceph-mon[53973]: pgmap v1394: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:33 vm10.local ceph-mon[53712]: pgmap v1394: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:34.660 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:34.661 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:34.687 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:34.688 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:34 vm11.local ceph-mon[53973]: pgmap v1395: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:35.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:34 vm10.local ceph-mon[53712]: pgmap v1395: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:37.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:37 vm10.local ceph-mon[53712]: pgmap v1396: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:37 vm11.local ceph-mon[53973]: pgmap v1396: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:39.689 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:39.690 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:39.717 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:39.718 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:39 vm10.local ceph-mon[53712]: pgmap v1397: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:39 vm11.local ceph-mon[53973]: pgmap v1397: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:41 vm11.local ceph-mon[53973]: pgmap v1398: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:41 vm10.local ceph-mon[53712]: pgmap v1398: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:43 vm11.local ceph-mon[53973]: pgmap v1399: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:43 vm10.local ceph-mon[53712]: pgmap v1399: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:44.719 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:44.720 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:44.745 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:44.745 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:44 vm11.local ceph-mon[53973]: pgmap v1400: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:44 vm10.local ceph-mon[53712]: pgmap v1400: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:47 vm10.local ceph-mon[53712]: pgmap v1401: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:47 vm11.local ceph-mon[53973]: pgmap v1401: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:49.746 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:49.747 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:49.773 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:49.774 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:49.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:49 vm10.local ceph-mon[53712]: pgmap v1402: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:49 vm11.local ceph-mon[53973]: pgmap v1402: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:51 vm11.local ceph-mon[53973]: pgmap v1403: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:51 vm10.local ceph-mon[53712]: pgmap v1403: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:52 vm11.local ceph-mon[53973]: pgmap v1404: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:52 vm10.local ceph-mon[53712]: pgmap v1404: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:54.775 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:54.775 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:54.802 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:54.802 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:12:56.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:55 vm11.local ceph-mon[53973]: pgmap v1405: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:56.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:55 vm10.local ceph-mon[53712]: pgmap v1405: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:57 vm11.local ceph-mon[53973]: pgmap v1406: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:57 vm10.local ceph-mon[53712]: pgmap v1406: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:12:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:12:58 vm11.local ceph-mon[53973]: pgmap v1407: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:12:58 vm10.local ceph-mon[53712]: pgmap v1407: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:12:59.804 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:12:59.804 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:12:59.853 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:12:59.853 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:01 vm10.local ceph-mon[53712]: pgmap v1408: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:01 vm11.local ceph-mon[53973]: pgmap v1408: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:03 vm10.local ceph-mon[53712]: pgmap v1409: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:03 vm11.local ceph-mon[53973]: pgmap v1409: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:04.854 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:04.855 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:04.881 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:04.881 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:04 vm11.local ceph-mon[53973]: pgmap v1410: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:05.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:04 vm10.local ceph-mon[53712]: pgmap v1410: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:07 vm10.local ceph-mon[53712]: pgmap v1411: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:07 vm11.local ceph-mon[53973]: pgmap v1411: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:09.883 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:09.883 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:09.909 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:09.909 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:09 vm10.local ceph-mon[53712]: pgmap v1412: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:09 vm11.local ceph-mon[53973]: pgmap v1412: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:11 vm10.local ceph-mon[53712]: pgmap v1413: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:11 vm11.local ceph-mon[53973]: pgmap v1413: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:13.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:13 vm10.local ceph-mon[53712]: pgmap v1414: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:13:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:13:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:13:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:13:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:13:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:13 vm11.local ceph-mon[53973]: pgmap v1414: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:13:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:13:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:13:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:13:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:13:14.911 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:14.911 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:14.968 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:14.969 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:14 vm11.local ceph-mon[53973]: pgmap v1415: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:14 vm10.local ceph-mon[53712]: pgmap v1415: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:17 vm10.local ceph-mon[53712]: pgmap v1416: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:17 vm11.local ceph-mon[53973]: pgmap v1416: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:19.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:19 vm10.local ceph-mon[53712]: pgmap v1417: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:19.970 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:19.971 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:19.999 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:19.999 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:19 vm11.local ceph-mon[53973]: pgmap v1417: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:21 vm10.local ceph-mon[53712]: pgmap v1418: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:21 vm11.local ceph-mon[53973]: pgmap v1418: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:23 vm10.local ceph-mon[53712]: pgmap v1419: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:23 vm11.local ceph-mon[53973]: pgmap v1419: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:25.000 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:25.001 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:25.027 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:25.028 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:24 vm11.local ceph-mon[53973]: pgmap v1420: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:24 vm10.local ceph-mon[53712]: pgmap v1420: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:27 vm10.local ceph-mon[53712]: pgmap v1421: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:27 vm11.local ceph-mon[53973]: pgmap v1421: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:13:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:13:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:13:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:13:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:29 vm10.local ceph-mon[53712]: pgmap v1422: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:30.029 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:30.030 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:30.056 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:30.056 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:29 vm11.local ceph-mon[53973]: pgmap v1422: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:31 vm10.local ceph-mon[53712]: pgmap v1423: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:31 vm11.local ceph-mon[53973]: pgmap v1423: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:33 vm10.local ceph-mon[53712]: pgmap v1424: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:33 vm11.local ceph-mon[53973]: pgmap v1424: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:35.057 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:35.058 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:35.084 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:35.084 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:34 vm11.local ceph-mon[53973]: pgmap v1425: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:34 vm10.local ceph-mon[53712]: pgmap v1425: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:37.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:37 vm10.local ceph-mon[53712]: pgmap v1426: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:37 vm11.local ceph-mon[53973]: pgmap v1426: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:39 vm10.local ceph-mon[53712]: pgmap v1427: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:40.086 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:40.086 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:39 vm11.local ceph-mon[53973]: pgmap v1427: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:40.111 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:40.112 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:41.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:41 vm10.local ceph-mon[53712]: pgmap v1428: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:41 vm11.local ceph-mon[53973]: pgmap v1428: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:43 vm10.local ceph-mon[53712]: pgmap v1429: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:43 vm11.local ceph-mon[53973]: pgmap v1429: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:44 vm11.local ceph-mon[53973]: pgmap v1430: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:45.113 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:45.114 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:45.163 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:45.164 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:44 vm10.local ceph-mon[53712]: pgmap v1430: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:47 vm10.local ceph-mon[53712]: pgmap v1431: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:47 vm11.local ceph-mon[53973]: pgmap v1431: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:49.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:49 vm10.local ceph-mon[53712]: pgmap v1432: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:49 vm11.local ceph-mon[53973]: pgmap v1432: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:50.166 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:50.167 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:50.193 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:50.193 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:51.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:51 vm10.local ceph-mon[53712]: pgmap v1433: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:51 vm11.local ceph-mon[53973]: pgmap v1433: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:53 vm10.local ceph-mon[53712]: pgmap v1434: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:53 vm11.local ceph-mon[53973]: pgmap v1434: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:54 vm11.local ceph-mon[53973]: pgmap v1435: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:54 vm10.local ceph-mon[53712]: pgmap v1435: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:13:55.195 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:13:55.195 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:13:55.222 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:13:55.223 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:13:57.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:57 vm10.local ceph-mon[53712]: pgmap v1436: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:57 vm11.local ceph-mon[53973]: pgmap v1436: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:13:59.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:13:59 vm10.local ceph-mon[53712]: pgmap v1437: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:13:59 vm11.local ceph-mon[53973]: pgmap v1437: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:00.224 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:00.225 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:00.250 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:00.251 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:01.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:01 vm10.local ceph-mon[53712]: pgmap v1438: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:01 vm11.local ceph-mon[53973]: pgmap v1438: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:03 vm10.local ceph-mon[53712]: pgmap v1439: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:03 vm11.local ceph-mon[53973]: pgmap v1439: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:04 vm11.local ceph-mon[53973]: pgmap v1440: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:05.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:04 vm10.local ceph-mon[53712]: pgmap v1440: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:05.252 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:05.253 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:05.279 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:05.279 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:07 vm10.local ceph-mon[53712]: pgmap v1441: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:07 vm11.local ceph-mon[53973]: pgmap v1441: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:09 vm10.local ceph-mon[53712]: pgmap v1442: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:09 vm11.local ceph-mon[53973]: pgmap v1442: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:10.281 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:10.281 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:10.309 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:10.310 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:11 vm10.local ceph-mon[53712]: pgmap v1443: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:11 vm11.local ceph-mon[53973]: pgmap v1443: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:13 vm10.local ceph-mon[53712]: pgmap v1444: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:14:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:14:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:14:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:14:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:14:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:13 vm11.local ceph-mon[53973]: pgmap v1444: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:14:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:14:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:14:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:14:14.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:14:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:14 vm11.local ceph-mon[53973]: pgmap v1445: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:14 vm10.local ceph-mon[53712]: pgmap v1445: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:15.312 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:15.312 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:15.364 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:15.365 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:17.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:17 vm10.local ceph-mon[53712]: pgmap v1446: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:17 vm11.local ceph-mon[53973]: pgmap v1446: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:19.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:19 vm10.local ceph-mon[53712]: pgmap v1447: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:19 vm11.local ceph-mon[53973]: pgmap v1447: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:20.366 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:20.367 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:20.394 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:20.395 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:21 vm10.local ceph-mon[53712]: pgmap v1448: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:21 vm11.local ceph-mon[53973]: pgmap v1448: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:23.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:23 vm10.local ceph-mon[53712]: pgmap v1449: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:23 vm11.local ceph-mon[53973]: pgmap v1449: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:24 vm11.local ceph-mon[53973]: pgmap v1450: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:25.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:24 vm10.local ceph-mon[53712]: pgmap v1450: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:25.396 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:25.397 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:25.423 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:25.423 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:27 vm10.local ceph-mon[53712]: pgmap v1451: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:27 vm11.local ceph-mon[53973]: pgmap v1451: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:14:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:14:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:14:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:14:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:29 vm10.local ceph-mon[53712]: pgmap v1452: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:29 vm11.local ceph-mon[53973]: pgmap v1452: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:30.425 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:30.425 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:30.453 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:30.453 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:31 vm10.local ceph-mon[53712]: pgmap v1453: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:31 vm11.local ceph-mon[53973]: pgmap v1453: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:33 vm10.local ceph-mon[53712]: pgmap v1454: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:33 vm11.local ceph-mon[53973]: pgmap v1454: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:34 vm11.local ceph-mon[53973]: pgmap v1455: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:34 vm10.local ceph-mon[53712]: pgmap v1455: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:35.454 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:35.455 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:35.481 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:35.481 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:37 vm10.local ceph-mon[53712]: pgmap v1456: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:37 vm11.local ceph-mon[53973]: pgmap v1456: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:39.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:39 vm10.local ceph-mon[53712]: pgmap v1457: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:39 vm11.local ceph-mon[53973]: pgmap v1457: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:40.483 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:40.483 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:40.509 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:40.509 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:41 vm10.local ceph-mon[53712]: pgmap v1458: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:41 vm11.local ceph-mon[53973]: pgmap v1458: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:43 vm10.local ceph-mon[53712]: pgmap v1459: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:44.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:43 vm11.local ceph-mon[53973]: pgmap v1459: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:44 vm11.local ceph-mon[53973]: pgmap v1460: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:44 vm10.local ceph-mon[53712]: pgmap v1460: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:45.511 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:45.511 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:45.539 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:45.540 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:47.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:47 vm10.local ceph-mon[53712]: pgmap v1461: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:48.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:47 vm11.local ceph-mon[53973]: pgmap v1461: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:49.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:49 vm10.local ceph-mon[53712]: pgmap v1462: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:50.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:49 vm11.local ceph-mon[53973]: pgmap v1462: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:50.541 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:50.542 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:50.569 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:50.570 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:51.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:51 vm10.local ceph-mon[53712]: pgmap v1463: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:52.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:51 vm11.local ceph-mon[53973]: pgmap v1463: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:53.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:53 vm10.local ceph-mon[53712]: pgmap v1464: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:53 vm11.local ceph-mon[53973]: pgmap v1464: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:55.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:54 vm11.local ceph-mon[53973]: pgmap v1465: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:55.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:54 vm10.local ceph-mon[53712]: pgmap v1465: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:14:55.571 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:14:55.572 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:14:55.600 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:14:55.601 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:14:58.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:57 vm11.local ceph-mon[53973]: pgmap v1466: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:14:58.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:57 vm10.local ceph-mon[53712]: pgmap v1466: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:00.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:14:59 vm11.local ceph-mon[53973]: pgmap v1467: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:14:59 vm10.local ceph-mon[53712]: pgmap v1467: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:00.603 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:00.603 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:00.636 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:00.636 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:01 vm11.local ceph-mon[53973]: pgmap v1468: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:01 vm10.local ceph-mon[53712]: pgmap v1468: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:02 vm11.local ceph-mon[53973]: pgmap v1469: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:03.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:02 vm10.local ceph-mon[53712]: pgmap v1469: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:05.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:04 vm11.local ceph-mon[53973]: pgmap v1470: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:05.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:04 vm10.local ceph-mon[53712]: pgmap v1470: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:05.638 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:05.638 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:05.672 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:05.673 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:07 vm11.local ceph-mon[53973]: pgmap v1471: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:07 vm10.local ceph-mon[53712]: pgmap v1471: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:08 vm11.local ceph-mon[53973]: pgmap v1472: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:08 vm10.local ceph-mon[53712]: pgmap v1472: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:10.675 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:10.675 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:10.701 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:10.701 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:11.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:11 vm10.local ceph-mon[53712]: pgmap v1473: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:11 vm11.local ceph-mon[53973]: pgmap v1473: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:13.758 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:13 vm10.local ceph-mon[53712]: pgmap v1474: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:13.758 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:15:13.758 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:15:13.758 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:13 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:15:13.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:13 vm11.local ceph-mon[53973]: pgmap v1474: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:13.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:15:13.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:15:13.841 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:13 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:15:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:15:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:15:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:14 vm11.local ceph-mon[53973]: pgmap v1475: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:15:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:15:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:14 vm10.local ceph-mon[53712]: pgmap v1475: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:15.703 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:15.703 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:15.732 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:15.733 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:17 vm10.local ceph-mon[53712]: pgmap v1476: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:17 vm11.local ceph-mon[53973]: pgmap v1476: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:19.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:19 vm10.local ceph-mon[53712]: pgmap v1477: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:19 vm11.local ceph-mon[53973]: pgmap v1477: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:20.734 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:20.735 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:20.761 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:20.762 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:21.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:21 vm10.local ceph-mon[53712]: pgmap v1478: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:21 vm11.local ceph-mon[53973]: pgmap v1478: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:23 vm10.local ceph-mon[53712]: pgmap v1479: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:23 vm11.local ceph-mon[53973]: pgmap v1479: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:24 vm11.local ceph-mon[53973]: pgmap v1480: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:24 vm10.local ceph-mon[53712]: pgmap v1480: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:25.763 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:25.764 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:25.790 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:25.790 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:27.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:27 vm10.local ceph-mon[53712]: pgmap v1481: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:28.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:27 vm11.local ceph-mon[53973]: pgmap v1481: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:15:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:15:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:15:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:15:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:29 vm10.local ceph-mon[53712]: pgmap v1482: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:30.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:29 vm11.local ceph-mon[53973]: pgmap v1482: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:30.792 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:30.792 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:30.818 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:30.819 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:31 vm10.local ceph-mon[53712]: pgmap v1483: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:31 vm11.local ceph-mon[53973]: pgmap v1483: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:33.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:33 vm11.local ceph-mon[53973]: pgmap v1484: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:33.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:33 vm10.local ceph-mon[53712]: pgmap v1484: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:34 vm11.local ceph-mon[53973]: pgmap v1485: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:35.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:34 vm10.local ceph-mon[53712]: pgmap v1485: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:35.820 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:35.821 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:35.849 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:35.849 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:37.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:37 vm10.local ceph-mon[53712]: pgmap v1486: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:38.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:37 vm11.local ceph-mon[53973]: pgmap v1486: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:39.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:39 vm10.local ceph-mon[53712]: pgmap v1487: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:39 vm11.local ceph-mon[53973]: pgmap v1487: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:40.851 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:40.851 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:40.876 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:40.877 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:41.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:41 vm10.local ceph-mon[53712]: pgmap v1488: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:41 vm11.local ceph-mon[53973]: pgmap v1488: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:43.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:42 vm10.local ceph-mon[53712]: pgmap v1489: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:43.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:42 vm11.local ceph-mon[53973]: pgmap v1489: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:44 vm11.local ceph-mon[53973]: pgmap v1490: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:44 vm10.local ceph-mon[53712]: pgmap v1490: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:45.878 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:45.878 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:45.906 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:45.907 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:47 vm10.local ceph-mon[53712]: pgmap v1491: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:47 vm11.local ceph-mon[53973]: pgmap v1491: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:49.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:49 vm10.local ceph-mon[53712]: pgmap v1492: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:49 vm11.local ceph-mon[53973]: pgmap v1492: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:50.908 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:50.909 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:50.937 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:50.938 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:51 vm10.local ceph-mon[53712]: pgmap v1493: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:51 vm11.local ceph-mon[53973]: pgmap v1493: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:53 vm11.local ceph-mon[53973]: pgmap v1494: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:54.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:53 vm10.local ceph-mon[53712]: pgmap v1494: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:54 vm11.local ceph-mon[53973]: pgmap v1495: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:54 vm10.local ceph-mon[53712]: pgmap v1495: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:55.940 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:15:55.940 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:15:55.966 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:15:55.966 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:15:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:57 vm11.local ceph-mon[53973]: pgmap v1496: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:57 vm10.local ceph-mon[53712]: pgmap v1496: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:15:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:15:58 vm11.local ceph-mon[53973]: pgmap v1497: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:15:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:15:58 vm10.local ceph-mon[53712]: pgmap v1497: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:00.967 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:00.968 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:00.995 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:00.995 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:01 vm10.local ceph-mon[53712]: pgmap v1498: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:01 vm11.local ceph-mon[53973]: pgmap v1498: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:03.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:03 vm10.local ceph-mon[53712]: pgmap v1499: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:03 vm11.local ceph-mon[53973]: pgmap v1499: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:04 vm11.local ceph-mon[53973]: pgmap v1500: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:04 vm10.local ceph-mon[53712]: pgmap v1500: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:05.997 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:05.998 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:06.023 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:06.024 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:07.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:07 vm10.local ceph-mon[53712]: pgmap v1501: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:07 vm11.local ceph-mon[53973]: pgmap v1501: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:09.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:09 vm10.local ceph-mon[53712]: pgmap v1502: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:09 vm11.local ceph-mon[53973]: pgmap v1502: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:11.025 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:11.026 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:11.051 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:11.051 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:11 vm10.local ceph-mon[53712]: pgmap v1503: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:11 vm11.local ceph-mon[53973]: pgmap v1503: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:13 vm10.local ceph-mon[53712]: pgmap v1504: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:13.985 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:13 vm11.local ceph-mon[53973]: pgmap v1504: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:14.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:16:14.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:16:14.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:16:14.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:16:14.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:16:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:16:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:16:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:16:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:16:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:16:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:15 vm10.local ceph-mon[53712]: pgmap v1505: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:16.053 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:16.053 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:16.080 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:16.081 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:16.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:15 vm11.local ceph-mon[53973]: pgmap v1505: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:17.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:17 vm10.local ceph-mon[53712]: pgmap v1506: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:17 vm11.local ceph-mon[53973]: pgmap v1506: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:19 vm11.local ceph-mon[53973]: pgmap v1507: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:19 vm10.local ceph-mon[53712]: pgmap v1507: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:21.083 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:21.083 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:20 vm11.local ceph-mon[53973]: pgmap v1508: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:21.110 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:21.111 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:20 vm10.local ceph-mon[53712]: pgmap v1508: 97 pgs: 97 active+clean; 453 KiB data, 67 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:23.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:23 vm10.local ceph-mon[53712]: pgmap v1509: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:23 vm11.local ceph-mon[53973]: pgmap v1509: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:24 vm10.local ceph-mon[53712]: pgmap v1510: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:24 vm11.local ceph-mon[53973]: pgmap v1510: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:26.112 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:26.113 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:26.142 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:26.143 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:27 vm10.local ceph-mon[53712]: pgmap v1511: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:28.044 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:27 vm11.local ceph-mon[53973]: pgmap v1511: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:16:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:16:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:16:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:16:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:29 vm10.local ceph-mon[53712]: pgmap v1512: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:29 vm11.local ceph-mon[53973]: pgmap v1512: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:31.144 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:31.144 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:31.172 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:31.173 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:31 vm11.local ceph-mon[53973]: pgmap v1513: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:31 vm10.local ceph-mon[53712]: pgmap v1513: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:33 vm11.local ceph-mon[53973]: pgmap v1514: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:33 vm10.local ceph-mon[53712]: pgmap v1514: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:34 vm11.local ceph-mon[53973]: pgmap v1515: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:35.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:34 vm10.local ceph-mon[53712]: pgmap v1515: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:36.174 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:36.175 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:36.262 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:36.262 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:37 vm11.local ceph-mon[53973]: pgmap v1516: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:37 vm10.local ceph-mon[53712]: pgmap v1516: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:39 vm11.local ceph-mon[53973]: pgmap v1517: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:39 vm10.local ceph-mon[53712]: pgmap v1517: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:40 vm11.local ceph-mon[53973]: pgmap v1518: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:41.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:40 vm10.local ceph-mon[53712]: pgmap v1518: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:41.264 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:41.264 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:41.291 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:41.291 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:43.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:43 vm10.local ceph-mon[53712]: pgmap v1519: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:43 vm11.local ceph-mon[53973]: pgmap v1519: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:44 vm11.local ceph-mon[53973]: pgmap v1520: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:45.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:44 vm10.local ceph-mon[53712]: pgmap v1520: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:46.293 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:46.294 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:46.340 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:46.341 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:47 vm10.local ceph-mon[53712]: pgmap v1521: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:47 vm11.local ceph-mon[53973]: pgmap v1521: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:49.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:49 vm10.local ceph-mon[53712]: pgmap v1522: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:49 vm11.local ceph-mon[53973]: pgmap v1522: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:51.342 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:51.343 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:51.368 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:51.369 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:51 vm10.local ceph-mon[53712]: pgmap v1523: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:51 vm11.local ceph-mon[53973]: pgmap v1523: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:53.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:53 vm10.local ceph-mon[53712]: pgmap v1524: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:54.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:53 vm11.local ceph-mon[53973]: pgmap v1524: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:54 vm11.local ceph-mon[53973]: pgmap v1525: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:54 vm10.local ceph-mon[53712]: pgmap v1525: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:16:56.371 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:16:56.372 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:16:56.399 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:16:56.399 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:16:57.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:57 vm10.local ceph-mon[53712]: pgmap v1526: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:57 vm11.local ceph-mon[53973]: pgmap v1526: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:16:59.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:16:59 vm10.local ceph-mon[53712]: pgmap v1527: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:16:59 vm11.local ceph-mon[53973]: pgmap v1527: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:01.401 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:01.401 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:01.427 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:01.428 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:01 vm10.local ceph-mon[53712]: pgmap v1528: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:01 vm11.local ceph-mon[53973]: pgmap v1528: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:03.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:03 vm10.local ceph-mon[53712]: pgmap v1529: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:03 vm11.local ceph-mon[53973]: pgmap v1529: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:04 vm11.local ceph-mon[53973]: pgmap v1530: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:17:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:04 vm10.local ceph-mon[53712]: pgmap v1530: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:17:06.429 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:06.430 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:06.455 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:06.456 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:07 vm11.local ceph-mon[53973]: pgmap v1531: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:07 vm10.local ceph-mon[53712]: pgmap v1531: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:09 vm11.local ceph-mon[53973]: pgmap v1532: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:17:10.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:09 vm10.local ceph-mon[53712]: pgmap v1532: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:17:11.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:10 vm11.local ceph-mon[53973]: pgmap v1533: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:17:11.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:10 vm10.local ceph-mon[53712]: pgmap v1533: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:17:11.457 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:11.458 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:11.524 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:11.524 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:13.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:13 vm10.local ceph-mon[53712]: pgmap v1534: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:14.029 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:13 vm11.local ceph-mon[53973]: pgmap v1534: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:17:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:17:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:17:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:17:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:17:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:17:15.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:14 vm11.local ceph-mon[53973]: pgmap v1535: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:17:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:17:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:17:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:17:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:17:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:17:15.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:14 vm10.local ceph-mon[53712]: pgmap v1535: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:16.526 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:16.527 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:16.554 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:16.554 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:17 vm11.local ceph-mon[53973]: pgmap v1536: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:17 vm10.local ceph-mon[53712]: pgmap v1536: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:19 vm11.local ceph-mon[53973]: pgmap v1537: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:20.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:19 vm10.local ceph-mon[53712]: pgmap v1537: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:21.555 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:21.556 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:21.584 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:21.584 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:21 vm11.local ceph-mon[53973]: pgmap v1538: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:21 vm10.local ceph-mon[53712]: pgmap v1538: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:22 vm11.local ceph-mon[53973]: pgmap v1539: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:22 vm10.local ceph-mon[53712]: pgmap v1539: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:24 vm11.local ceph-mon[53973]: pgmap v1540: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:24 vm10.local ceph-mon[53712]: pgmap v1540: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:26.586 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:26.586 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:26.613 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:26.614 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:27 vm10.local ceph-mon[53712]: pgmap v1541: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:27 vm11.local ceph-mon[53973]: pgmap v1541: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:17:29.219 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:17:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:17:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:17:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:29 vm10.local ceph-mon[53712]: pgmap v1542: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:30.242 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:29 vm11.local ceph-mon[53973]: pgmap v1542: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:31.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:30 vm10.local ceph-mon[53712]: pgmap v1543: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:30 vm11.local ceph-mon[53973]: pgmap v1543: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:31.615 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:31.616 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:31.643 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:31.643 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:33 vm10.local ceph-mon[53712]: pgmap v1544: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:33 vm11.local ceph-mon[53973]: pgmap v1544: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:34 vm11.local ceph-mon[53973]: pgmap v1545: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:34 vm10.local ceph-mon[53712]: pgmap v1545: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:36.645 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:36.645 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:36.675 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:36.675 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:37 vm11.local ceph-mon[53973]: pgmap v1546: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:37 vm10.local ceph-mon[53712]: pgmap v1546: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:38 vm11.local ceph-mon[53973]: pgmap v1547: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:39.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:38 vm10.local ceph-mon[53712]: pgmap v1547: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:41.677 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:41.677 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:41.744 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:41.745 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:41.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:41 vm10.local ceph-mon[53712]: pgmap v1548: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:41 vm11.local ceph-mon[53973]: pgmap v1548: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:43 vm11.local ceph-mon[53973]: pgmap v1549: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:43 vm10.local ceph-mon[53712]: pgmap v1549: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:44 vm11.local ceph-mon[53973]: pgmap v1550: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:44 vm10.local ceph-mon[53712]: pgmap v1550: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:46.746 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:46.747 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:46.780 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:46.780 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:47 vm10.local ceph-mon[53712]: pgmap v1551: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:47 vm11.local ceph-mon[53973]: pgmap v1551: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:49.854 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:49 vm10.local ceph-mon[53712]: pgmap v1552: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:49 vm11.local ceph-mon[53973]: pgmap v1552: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:51.782 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:51.783 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:51.809 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:51.810 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:51 vm10.local ceph-mon[53712]: pgmap v1553: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:51 vm11.local ceph-mon[53973]: pgmap v1553: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:53 vm11.local ceph-mon[53973]: pgmap v1554: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:53 vm10.local ceph-mon[53712]: pgmap v1554: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:54 vm11.local ceph-mon[53973]: pgmap v1555: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:54 vm10.local ceph-mon[53712]: pgmap v1555: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:17:56.811 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:17:56.812 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:17:56.840 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:17:56.841 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:17:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:56 vm11.local ceph-mon[53973]: pgmap v1556: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:56 vm10.local ceph-mon[53712]: pgmap v1556: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:17:59.906 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:17:59 vm10.local ceph-mon[53712]: pgmap v1557: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:17:59 vm11.local ceph-mon[53973]: pgmap v1557: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:01.843 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:01.843 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:01.870 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:01.871 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:01.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:01 vm10.local ceph-mon[53712]: pgmap v1558: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:01 vm11.local ceph-mon[53973]: pgmap v1558: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:03.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:03 vm10.local ceph-mon[53712]: pgmap v1559: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:03 vm11.local ceph-mon[53973]: pgmap v1559: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:04 vm11.local ceph-mon[53973]: pgmap v1560: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:05.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:04 vm10.local ceph-mon[53712]: pgmap v1560: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:06.872 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:06.873 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:06.904 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:06.905 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:07 vm11.local ceph-mon[53973]: pgmap v1561: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:07 vm10.local ceph-mon[53712]: pgmap v1561: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:08 vm11.local ceph-mon[53973]: pgmap v1562: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:08 vm10.local ceph-mon[53712]: pgmap v1562: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:11.907 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:11.907 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:11 vm10.local ceph-mon[53712]: pgmap v1563: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:11.935 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:11.936 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:11 vm11.local ceph-mon[53973]: pgmap v1563: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:12 vm11.local ceph-mon[53973]: pgmap v1564: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:13.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:12 vm10.local ceph-mon[53712]: pgmap v1564: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:14 vm11.local ceph-mon[53973]: pgmap v1565: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:14 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:14 vm10.local ceph-mon[53712]: pgmap v1565: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:14.815 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:14 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:18:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:16 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:18:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:16 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:18:16.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:16 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:18:16.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:16 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:18:16.938 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:16.938 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:16.966 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:16.967 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:17 vm11.local ceph-mon[53973]: pgmap v1566: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:17 vm10.local ceph-mon[53712]: pgmap v1566: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:19 vm10.local ceph-mon[53712]: pgmap v1567: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:19 vm11.local ceph-mon[53973]: pgmap v1567: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:21 vm10.local ceph-mon[53712]: pgmap v1568: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:21.968 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:21.969 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:21.998 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:21.998 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:21 vm11.local ceph-mon[53973]: pgmap v1568: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:23 vm11.local ceph-mon[53973]: pgmap v1569: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:23 vm10.local ceph-mon[53712]: pgmap v1569: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:24 vm11.local ceph-mon[53973]: pgmap v1570: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:24 vm10.local ceph-mon[53712]: pgmap v1570: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:26.999 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:27.000 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:27.026 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:27.027 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:27.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:27 vm10.local ceph-mon[53712]: pgmap v1571: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:28.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:27 vm11.local ceph-mon[53973]: pgmap v1571: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:28.938 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:18:28.938 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:18:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:18:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:18:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:29 vm11.local ceph-mon[53973]: pgmap v1572: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:29 vm10.local ceph-mon[53712]: pgmap v1572: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:31 vm11.local ceph-mon[53973]: pgmap v1573: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:31.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:30 vm10.local ceph-mon[53712]: pgmap v1573: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:32.028 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:32.029 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:32.055 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:32.056 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:33 vm10.local ceph-mon[53712]: pgmap v1574: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:33 vm11.local ceph-mon[53973]: pgmap v1574: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:34 vm10.local ceph-mon[53712]: pgmap v1575: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:34 vm11.local ceph-mon[53973]: pgmap v1575: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:37.057 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:37.057 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:37.083 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:37.084 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:37.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:37 vm10.local ceph-mon[53712]: pgmap v1576: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:37 vm11.local ceph-mon[53973]: pgmap v1576: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:39.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:39 vm10.local ceph-mon[53712]: pgmap v1577: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:39 vm11.local ceph-mon[53973]: pgmap v1577: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:42.085 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:42.086 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:41 vm11.local ceph-mon[53973]: pgmap v1578: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:42.114 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:42.114 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:42.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:41 vm10.local ceph-mon[53712]: pgmap v1578: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:44.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:43 vm11.local ceph-mon[53973]: pgmap v1579: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:44.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:43 vm10.local ceph-mon[53712]: pgmap v1579: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:44 vm10.local ceph-mon[53712]: pgmap v1580: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:44 vm11.local ceph-mon[53973]: pgmap v1580: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:47.116 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:47.116 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:47.144 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:47.145 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:47.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:47 vm10.local ceph-mon[53712]: pgmap v1581: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:47 vm11.local ceph-mon[53973]: pgmap v1581: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:49 vm11.local ceph-mon[53973]: pgmap v1582: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:49 vm10.local ceph-mon[53712]: pgmap v1582: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:51.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:50 vm11.local ceph-mon[53973]: pgmap v1583: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:50 vm10.local ceph-mon[53712]: pgmap v1583: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:52.147 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:52.147 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:52.173 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:52.174 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:53.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:53 vm10.local ceph-mon[53712]: pgmap v1584: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:53 vm11.local ceph-mon[53973]: pgmap v1584: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:55 vm11.local ceph-mon[53973]: pgmap v1585: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:55 vm10.local ceph-mon[53712]: pgmap v1585: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:18:57.175 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:18:57.176 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:18:57.203 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:18:57.204 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:18:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:57 vm11.local ceph-mon[53973]: pgmap v1586: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:18:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:57 vm10.local ceph-mon[53712]: pgmap v1586: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:18:59 vm11.local ceph-mon[53973]: pgmap v1587: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:18:59 vm10.local ceph-mon[53712]: pgmap v1587: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:02.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:01 vm11.local ceph-mon[53973]: pgmap v1588: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:01 vm10.local ceph-mon[53712]: pgmap v1588: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:02.205 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:02.206 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:02.234 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:02.235 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:03.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:02 vm11.local ceph-mon[53973]: pgmap v1589: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:03.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:02 vm10.local ceph-mon[53712]: pgmap v1589: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:04 vm11.local ceph-mon[53973]: pgmap v1590: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:05.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:04 vm10.local ceph-mon[53712]: pgmap v1590: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:07.236 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:07.237 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:07.266 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:07.266 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:07 vm11.local ceph-mon[53973]: pgmap v1591: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:07 vm10.local ceph-mon[53712]: pgmap v1591: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:09.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:08 vm10.local ceph-mon[53712]: pgmap v1592: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:08 vm11.local ceph-mon[53973]: pgmap v1592: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:11 vm11.local ceph-mon[53973]: pgmap v1593: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:11 vm10.local ceph-mon[53712]: pgmap v1593: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:12.267 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:12.268 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:12.294 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:12.294 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:13 vm11.local ceph-mon[53973]: pgmap v1594: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:13 vm10.local ceph-mon[53712]: pgmap v1594: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:15.266 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:14 vm11.local ceph-mon[53973]: pgmap v1595: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:15.272 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:14 vm10.local ceph-mon[53712]: pgmap v1595: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:15 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:19:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:15 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:19:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:15 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:19:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:15 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:19:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:15 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:19:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:15 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:19:16.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:15 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:19:16.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:15 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:19:16.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:15 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:19:16.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:15 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:19:16.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:15 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:19:16.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:15 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:19:17.296 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:17.296 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:17.337 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:17.338 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:16 vm11.local ceph-mon[53973]: pgmap v1596: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:17.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:16 vm10.local ceph-mon[53712]: pgmap v1596: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:19.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:19 vm10.local ceph-mon[53712]: pgmap v1597: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:19 vm11.local ceph-mon[53973]: pgmap v1597: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:21 vm11.local ceph-mon[53973]: pgmap v1598: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:21 vm10.local ceph-mon[53712]: pgmap v1598: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:22.339 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:22.340 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:22.367 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:22.367 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:23 vm11.local ceph-mon[53973]: pgmap v1599: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:23 vm10.local ceph-mon[53712]: pgmap v1599: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:24 vm11.local ceph-mon[53973]: pgmap v1600: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:25.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:24 vm10.local ceph-mon[53712]: pgmap v1600: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:27.368 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:27.369 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:27.395 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:27.396 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:27 vm11.local ceph-mon[53973]: pgmap v1601: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:27 vm10.local ceph-mon[53712]: pgmap v1601: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:19:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:19:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:19:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:19:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:29 vm10.local ceph-mon[53712]: pgmap v1602: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:29 vm11.local ceph-mon[53973]: pgmap v1602: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:31 vm11.local ceph-mon[53973]: pgmap v1603: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:32.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:31 vm10.local ceph-mon[53712]: pgmap v1603: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:32.397 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:32.398 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:32.424 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:32.425 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:33 vm11.local ceph-mon[53973]: pgmap v1604: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:33 vm10.local ceph-mon[53712]: pgmap v1604: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:34 vm11.local ceph-mon[53973]: pgmap v1605: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:34 vm10.local ceph-mon[53712]: pgmap v1605: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:37.426 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:37.427 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:37.454 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:37.455 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:37 vm11.local ceph-mon[53973]: pgmap v1606: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:37 vm10.local ceph-mon[53712]: pgmap v1606: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:39 vm11.local ceph-mon[53973]: pgmap v1607: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:39 vm10.local ceph-mon[53712]: pgmap v1607: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:41 vm11.local ceph-mon[53973]: pgmap v1608: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:42.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:41 vm10.local ceph-mon[53712]: pgmap v1608: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:42.456 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:42.457 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:42.487 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:42.487 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:44.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:43 vm11.local ceph-mon[53973]: pgmap v1609: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:43 vm10.local ceph-mon[53712]: pgmap v1609: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:44 vm11.local ceph-mon[53973]: pgmap v1610: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:45.420 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:44 vm10.local ceph-mon[53712]: pgmap v1610: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:47.489 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:47.489 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:47.515 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:47.516 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:47 vm11.local ceph-mon[53973]: pgmap v1611: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:47 vm10.local ceph-mon[53712]: pgmap v1611: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:49 vm11.local ceph-mon[53973]: pgmap v1612: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:50.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:49 vm10.local ceph-mon[53712]: pgmap v1612: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:51 vm11.local ceph-mon[53973]: pgmap v1613: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:51 vm10.local ceph-mon[53712]: pgmap v1613: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:52.517 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:52.518 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:52.545 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:52.546 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:53 vm11.local ceph-mon[53973]: pgmap v1614: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:54.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:53 vm10.local ceph-mon[53712]: pgmap v1614: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:55.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:55 vm11.local ceph-mon[53973]: pgmap v1615: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:55.438 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:55 vm10.local ceph-mon[53712]: pgmap v1615: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:19:57.547 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:19:57.547 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:19:57.576 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:19:57.577 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:19:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:57 vm11.local ceph-mon[53973]: pgmap v1616: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:19:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:57 vm10.local ceph-mon[53712]: pgmap v1616: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:19:59 vm11.local ceph-mon[53973]: pgmap v1617: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:19:59 vm10.local ceph-mon[53712]: pgmap v1617: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:00 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T18:20:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:00 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T18:20:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:01 vm11.local ceph-mon[53973]: pgmap v1618: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:01 vm10.local ceph-mon[53712]: pgmap v1618: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:02.579 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:02.579 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:02.606 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:02.607 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:03 vm11.local ceph-mon[53973]: pgmap v1619: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:04.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:03 vm10.local ceph-mon[53712]: pgmap v1619: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:05 vm11.local ceph-mon[53973]: pgmap v1620: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:05 vm10.local ceph-mon[53712]: pgmap v1620: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:07.608 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:07.609 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:07.636 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:07.637 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:07 vm11.local ceph-mon[53973]: pgmap v1621: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:07 vm10.local ceph-mon[53712]: pgmap v1621: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:09 vm11.local ceph-mon[53973]: pgmap v1622: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:09 vm10.local ceph-mon[53712]: pgmap v1622: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:11 vm11.local ceph-mon[53973]: pgmap v1623: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:11 vm10.local ceph-mon[53712]: pgmap v1623: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:12.638 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:12.639 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:12.666 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:12.667 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:13 vm11.local ceph-mon[53973]: pgmap v1624: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:13 vm10.local ceph-mon[53712]: pgmap v1624: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:15.316 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:15 vm11.local ceph-mon[53973]: pgmap v1625: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:15.324 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:15 vm10.local ceph-mon[53712]: pgmap v1625: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:16.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:16 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:20:16.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:16 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:20:16.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:16 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:20:16.236 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:16 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:20:16.237 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:16 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:20:16.237 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:16 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:20:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:20:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:20:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:17 vm11.local ceph-mon[53973]: pgmap v1626: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:20:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:20:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:20:17.668 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:17.669 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:20:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:20:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:17 vm10.local ceph-mon[53712]: pgmap v1626: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:20:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:20:17.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:20:17.697 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:17.698 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:19 vm11.local ceph-mon[53973]: pgmap v1627: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:19 vm10.local ceph-mon[53712]: pgmap v1627: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:21 vm11.local ceph-mon[53973]: pgmap v1628: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:22.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:21 vm10.local ceph-mon[53712]: pgmap v1628: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:22.700 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:22.700 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:22.727 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:22.728 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:23 vm11.local ceph-mon[53973]: pgmap v1629: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:23 vm10.local ceph-mon[53712]: pgmap v1629: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:25 vm11.local ceph-mon[53973]: pgmap v1630: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:25 vm10.local ceph-mon[53712]: pgmap v1630: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:27.729 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:27.730 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:27.767 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:27.768 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:27 vm11.local ceph-mon[53973]: pgmap v1631: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:27 vm10.local ceph-mon[53712]: pgmap v1631: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:20:29.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:20:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:20:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:20:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:29 vm11.local ceph-mon[53973]: pgmap v1632: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:29 vm10.local ceph-mon[53712]: pgmap v1632: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:31 vm11.local ceph-mon[53973]: pgmap v1633: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:32.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:31 vm10.local ceph-mon[53712]: pgmap v1633: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:32.770 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:32.770 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:32.799 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:32.800 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:33 vm11.local ceph-mon[53973]: pgmap v1634: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:33 vm10.local ceph-mon[53712]: pgmap v1634: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:35 vm11.local ceph-mon[53973]: pgmap v1635: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:35 vm10.local ceph-mon[53712]: pgmap v1635: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:37.801 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:37.802 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:37.832 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:37.832 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:37 vm11.local ceph-mon[53973]: pgmap v1636: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:37 vm10.local ceph-mon[53712]: pgmap v1636: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:39 vm11.local ceph-mon[53973]: pgmap v1637: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:39 vm10.local ceph-mon[53712]: pgmap v1637: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:41 vm11.local ceph-mon[53973]: pgmap v1638: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:41 vm10.local ceph-mon[53712]: pgmap v1638: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:42.834 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:42.835 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:42.862 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:42.863 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:43 vm11.local ceph-mon[53973]: pgmap v1639: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:43 vm10.local ceph-mon[53712]: pgmap v1639: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:45 vm11.local ceph-mon[53973]: pgmap v1640: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:45 vm10.local ceph-mon[53712]: pgmap v1640: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:47.865 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:47.866 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:47.895 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:47.895 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:47 vm11.local ceph-mon[53973]: pgmap v1641: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:47 vm10.local ceph-mon[53712]: pgmap v1641: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:49 vm11.local ceph-mon[53973]: pgmap v1642: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:49 vm10.local ceph-mon[53712]: pgmap v1642: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:51 vm11.local ceph-mon[53973]: pgmap v1643: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:51 vm10.local ceph-mon[53712]: pgmap v1643: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:52.896 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:52.897 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:52.922 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:52.923 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:53 vm11.local ceph-mon[53973]: pgmap v1644: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:53 vm10.local ceph-mon[53712]: pgmap v1644: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:55 vm11.local ceph-mon[53973]: pgmap v1645: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:55 vm10.local ceph-mon[53712]: pgmap v1645: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:20:57.924 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:20:57.924 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:20:57.950 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:20:57.951 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:20:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:57 vm11.local ceph-mon[53973]: pgmap v1646: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:20:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:57 vm10.local ceph-mon[53712]: pgmap v1646: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:20:59 vm11.local ceph-mon[53973]: pgmap v1647: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:20:59 vm10.local ceph-mon[53712]: pgmap v1647: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:01 vm11.local ceph-mon[53973]: pgmap v1648: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:01 vm10.local ceph-mon[53712]: pgmap v1648: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:02.952 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:02.952 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:02.978 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:02.979 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:03 vm11.local ceph-mon[53973]: pgmap v1649: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:03 vm10.local ceph-mon[53712]: pgmap v1649: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:05 vm11.local ceph-mon[53973]: pgmap v1650: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:05 vm10.local ceph-mon[53712]: pgmap v1650: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:07 vm11.local ceph-mon[53973]: pgmap v1651: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:07 vm10.local ceph-mon[53712]: pgmap v1651: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:07.980 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:07.980 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:08.006 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:08.006 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:09 vm11.local ceph-mon[53973]: pgmap v1652: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:09 vm10.local ceph-mon[53712]: pgmap v1652: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:11 vm11.local ceph-mon[53973]: pgmap v1653: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:11 vm10.local ceph-mon[53712]: pgmap v1653: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:13.007 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:13.008 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:13.083 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:13.084 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:14.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:13 vm10.local ceph-mon[53712]: pgmap v1654: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:14 vm11.local ceph-mon[53973]: pgmap v1654: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:15 vm10.local ceph-mon[53712]: pgmap v1655: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:15 vm11.local ceph-mon[53973]: pgmap v1655: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:18.085 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:18.086 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:17 vm11.local ceph-mon[53973]: pgmap v1656: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:21:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:21:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:21:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:21:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:21:18.113 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:18.113 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:17 vm10.local ceph-mon[53712]: pgmap v1656: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:21:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:21:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:21:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:21:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:21:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:19 vm11.local ceph-mon[53973]: pgmap v1657: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:19 vm10.local ceph-mon[53712]: pgmap v1657: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:20 vm10.local ceph-mon[53712]: pgmap v1658: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:20 vm11.local ceph-mon[53973]: pgmap v1658: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:23.114 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:23.115 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:23.142 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:23.143 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:23 vm11.local ceph-mon[53973]: pgmap v1659: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:23 vm10.local ceph-mon[53712]: pgmap v1659: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:25 vm11.local ceph-mon[53973]: pgmap v1660: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:25 vm10.local ceph-mon[53712]: pgmap v1660: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:27 vm11.local ceph-mon[53973]: pgmap v1661: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:28.144 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:28.145 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:28.172 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:28.172 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:27 vm10.local ceph-mon[53712]: pgmap v1661: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:21:29.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:21:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:21:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:21:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:29 vm11.local ceph-mon[53973]: pgmap v1662: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:29 vm10.local ceph-mon[53712]: pgmap v1662: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:31 vm11.local ceph-mon[53973]: pgmap v1663: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:31 vm10.local ceph-mon[53712]: pgmap v1663: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:33.173 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:33.174 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:33.200 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:33.201 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:33 vm11.local ceph-mon[53973]: pgmap v1664: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:33 vm10.local ceph-mon[53712]: pgmap v1664: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:35 vm11.local ceph-mon[53973]: pgmap v1665: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:35 vm10.local ceph-mon[53712]: pgmap v1665: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:37 vm11.local ceph-mon[53973]: pgmap v1666: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:37 vm10.local ceph-mon[53712]: pgmap v1666: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:38.202 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:38.203 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:38.228 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:38.229 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:39 vm11.local ceph-mon[53973]: pgmap v1667: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:39 vm10.local ceph-mon[53712]: pgmap v1667: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:41 vm11.local ceph-mon[53973]: pgmap v1668: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:41 vm10.local ceph-mon[53712]: pgmap v1668: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:43.230 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:43.231 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:43.291 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:43.291 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:43 vm11.local ceph-mon[53973]: pgmap v1669: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:43 vm10.local ceph-mon[53712]: pgmap v1669: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:45 vm11.local ceph-mon[53973]: pgmap v1670: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:45 vm10.local ceph-mon[53712]: pgmap v1670: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:47 vm11.local ceph-mon[53973]: pgmap v1671: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:47 vm10.local ceph-mon[53712]: pgmap v1671: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:48.293 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:48.293 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:48.319 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:48.320 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:49 vm11.local ceph-mon[53973]: pgmap v1672: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:49 vm10.local ceph-mon[53712]: pgmap v1672: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:51 vm11.local ceph-mon[53973]: pgmap v1673: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:51 vm10.local ceph-mon[53712]: pgmap v1673: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:53.321 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:53.322 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:53.348 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:53.349 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:21:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:53 vm11.local ceph-mon[53973]: pgmap v1674: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:54.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:53 vm10.local ceph-mon[53712]: pgmap v1674: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:55 vm10.local ceph-mon[53712]: pgmap v1675: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:55 vm11.local ceph-mon[53973]: pgmap v1675: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:21:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:57 vm11.local ceph-mon[53973]: pgmap v1676: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:58.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:57 vm10.local ceph-mon[53712]: pgmap v1676: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:21:58.350 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:21:58.359 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:21:58.392 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:21:58.393 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:21:59 vm11.local ceph-mon[53973]: pgmap v1677: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:21:59 vm10.local ceph-mon[53712]: pgmap v1677: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:01 vm11.local ceph-mon[53973]: pgmap v1678: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:01 vm10.local ceph-mon[53712]: pgmap v1678: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:03.394 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:03.395 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:03.422 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:03.423 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:03 vm11.local ceph-mon[53973]: pgmap v1679: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:03 vm10.local ceph-mon[53712]: pgmap v1679: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:05 vm11.local ceph-mon[53973]: pgmap v1680: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:05 vm10.local ceph-mon[53712]: pgmap v1680: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:07 vm11.local ceph-mon[53973]: pgmap v1681: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:07 vm10.local ceph-mon[53712]: pgmap v1681: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:08.424 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:08.425 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:08.480 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:08.480 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:09 vm11.local ceph-mon[53973]: pgmap v1682: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:09 vm10.local ceph-mon[53712]: pgmap v1682: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:11 vm11.local ceph-mon[53973]: pgmap v1683: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:11 vm10.local ceph-mon[53712]: pgmap v1683: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:13.482 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:13.483 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:13.510 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:13.510 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:13 vm11.local ceph-mon[53973]: pgmap v1684: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:13 vm10.local ceph-mon[53712]: pgmap v1684: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:15 vm11.local ceph-mon[53973]: pgmap v1685: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:15 vm10.local ceph-mon[53712]: pgmap v1685: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:17 vm11.local ceph-mon[53973]: pgmap v1686: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:22:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:22:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:22:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:22:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:22:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:17 vm10.local ceph-mon[53712]: pgmap v1686: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:18.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:22:18.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:22:18.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:22:18.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:22:18.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:22:18.512 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:18.512 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:18.539 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:18.540 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:19 vm11.local ceph-mon[53973]: pgmap v1687: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:20.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:19 vm10.local ceph-mon[53712]: pgmap v1687: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:21 vm11.local ceph-mon[53973]: pgmap v1688: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:21 vm10.local ceph-mon[53712]: pgmap v1688: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:23.541 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:23.542 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:23.569 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:23.570 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:23 vm11.local ceph-mon[53973]: pgmap v1689: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:23 vm10.local ceph-mon[53712]: pgmap v1689: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:25 vm11.local ceph-mon[53973]: pgmap v1690: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:25 vm10.local ceph-mon[53712]: pgmap v1690: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:27 vm11.local ceph-mon[53973]: pgmap v1691: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:27 vm10.local ceph-mon[53712]: pgmap v1691: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:28.572 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:28.572 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:28.600 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:28.601 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:22:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:22:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:22:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:22:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:29 vm11.local ceph-mon[53973]: pgmap v1692: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:30.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:29 vm10.local ceph-mon[53712]: pgmap v1692: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:31 vm11.local ceph-mon[53973]: pgmap v1693: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:31 vm10.local ceph-mon[53712]: pgmap v1693: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:33.603 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:33.603 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:33.638 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:33.638 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:33 vm11.local ceph-mon[53973]: pgmap v1694: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:33 vm10.local ceph-mon[53712]: pgmap v1694: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:35 vm11.local ceph-mon[53973]: pgmap v1695: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:35 vm10.local ceph-mon[53712]: pgmap v1695: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:37 vm11.local ceph-mon[53973]: pgmap v1696: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:37 vm10.local ceph-mon[53712]: pgmap v1696: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:38.640 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:38.641 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:38.668 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:38.669 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:39 vm11.local ceph-mon[53973]: pgmap v1697: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:39 vm10.local ceph-mon[53712]: pgmap v1697: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:41 vm11.local ceph-mon[53973]: pgmap v1698: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:41 vm10.local ceph-mon[53712]: pgmap v1698: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:43.670 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:43.671 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:43.705 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:43.705 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:43 vm11.local ceph-mon[53973]: pgmap v1699: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:43 vm10.local ceph-mon[53712]: pgmap v1699: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:45.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:45 vm10.local ceph-mon[53712]: pgmap v1700: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:45.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:45 vm11.local ceph-mon[53973]: pgmap v1700: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:47 vm11.local ceph-mon[53973]: pgmap v1701: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:47 vm10.local ceph-mon[53712]: pgmap v1701: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:48.706 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:48.707 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:48.734 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:48.735 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:49 vm11.local ceph-mon[53973]: pgmap v1702: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:49 vm10.local ceph-mon[53712]: pgmap v1702: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:51 vm11.local ceph-mon[53973]: pgmap v1703: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:52.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:51 vm10.local ceph-mon[53712]: pgmap v1703: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:53.737 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:53.737 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:53.773 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:53.773 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:22:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:53 vm11.local ceph-mon[53973]: pgmap v1704: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:53 vm10.local ceph-mon[53712]: pgmap v1704: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:55 vm11.local ceph-mon[53973]: pgmap v1705: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:55 vm10.local ceph-mon[53712]: pgmap v1705: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:22:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:57 vm11.local ceph-mon[53973]: pgmap v1706: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:57 vm10.local ceph-mon[53712]: pgmap v1706: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:22:58.775 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:22:58.776 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:22:58.808 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:22:58.809 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:22:59 vm11.local ceph-mon[53973]: pgmap v1707: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:22:59 vm10.local ceph-mon[53712]: pgmap v1707: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:01 vm11.local ceph-mon[53973]: pgmap v1708: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:01 vm10.local ceph-mon[53712]: pgmap v1708: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:03.810 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:03.810 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:03.837 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:03.838 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:03 vm11.local ceph-mon[53973]: pgmap v1709: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:03 vm10.local ceph-mon[53712]: pgmap v1709: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:05 vm11.local ceph-mon[53973]: pgmap v1710: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:05 vm10.local ceph-mon[53712]: pgmap v1710: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:07 vm11.local ceph-mon[53973]: pgmap v1711: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:07 vm10.local ceph-mon[53712]: pgmap v1711: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:08.839 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:08.839 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:08.866 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:08.867 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:09 vm11.local ceph-mon[53973]: pgmap v1712: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:09 vm10.local ceph-mon[53712]: pgmap v1712: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:11 vm11.local ceph-mon[53973]: pgmap v1713: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:11 vm10.local ceph-mon[53712]: pgmap v1713: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:13.868 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:13.869 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:13.895 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:13.896 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:13 vm11.local ceph-mon[53973]: pgmap v1714: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:13 vm10.local ceph-mon[53712]: pgmap v1714: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:15 vm10.local ceph-mon[53712]: pgmap v1715: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:15 vm11.local ceph-mon[53973]: pgmap v1715: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:17.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:17 vm10.local ceph-mon[53712]: pgmap v1716: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:17.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:23:17.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:23:17.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:17 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:23:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:17 vm11.local ceph-mon[53973]: pgmap v1716: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:23:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:23:18.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:17 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:23:18.897 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:18.898 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:18.924 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:18.924 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:23:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:23:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:18 vm10.local ceph-mon[53712]: pgmap v1717: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:23:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:23:19.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:18 vm11.local ceph-mon[53973]: pgmap v1717: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:21 vm11.local ceph-mon[53973]: pgmap v1718: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:22.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:21 vm10.local ceph-mon[53712]: pgmap v1718: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:23.926 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:23.926 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:23.954 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:23.954 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:23 vm11.local ceph-mon[53973]: pgmap v1719: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:23 vm10.local ceph-mon[53712]: pgmap v1719: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:25 vm10.local ceph-mon[53712]: pgmap v1720: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:25 vm11.local ceph-mon[53973]: pgmap v1720: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:27 vm11.local ceph-mon[53973]: pgmap v1721: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:28.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:27 vm10.local ceph-mon[53712]: pgmap v1721: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:28.956 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:28.957 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:28.992 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:28.992 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:23:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:23:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:23:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:23:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:29 vm11.local ceph-mon[53973]: pgmap v1722: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:29 vm10.local ceph-mon[53712]: pgmap v1722: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:31 vm11.local ceph-mon[53973]: pgmap v1723: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:31 vm10.local ceph-mon[53712]: pgmap v1723: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:33.994 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:33.994 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:34.021 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:34.021 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:33 vm11.local ceph-mon[53973]: pgmap v1724: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:33 vm10.local ceph-mon[53712]: pgmap v1724: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:35 vm10.local ceph-mon[53712]: pgmap v1725: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:35 vm11.local ceph-mon[53973]: pgmap v1725: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:37 vm11.local ceph-mon[53973]: pgmap v1726: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:37 vm10.local ceph-mon[53712]: pgmap v1726: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:39.023 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:39.023 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:39.050 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:39.050 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:39 vm11.local ceph-mon[53973]: pgmap v1727: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:39 vm10.local ceph-mon[53712]: pgmap v1727: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:41 vm11.local ceph-mon[53973]: pgmap v1728: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:42.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:41 vm10.local ceph-mon[53712]: pgmap v1728: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:44.052 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:44.052 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:44.079 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:44.079 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:43 vm11.local ceph-mon[53973]: pgmap v1729: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:44.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:43 vm10.local ceph-mon[53712]: pgmap v1729: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:45 vm10.local ceph-mon[53712]: pgmap v1730: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:45 vm11.local ceph-mon[53973]: pgmap v1730: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:47 vm11.local ceph-mon[53973]: pgmap v1731: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:47 vm10.local ceph-mon[53712]: pgmap v1731: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:49.081 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:49.081 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:49.114 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:49.115 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:49 vm11.local ceph-mon[53973]: pgmap v1732: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:49 vm10.local ceph-mon[53712]: pgmap v1732: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:51 vm11.local ceph-mon[53973]: pgmap v1733: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:51 vm10.local ceph-mon[53712]: pgmap v1733: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:53 vm11.local ceph-mon[53973]: pgmap v1734: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:54.116 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:54.117 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:54.144 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:54.144 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:23:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:53 vm10.local ceph-mon[53712]: pgmap v1734: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:55 vm10.local ceph-mon[53712]: pgmap v1735: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:55 vm11.local ceph-mon[53973]: pgmap v1735: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:23:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:57 vm11.local ceph-mon[53973]: pgmap v1736: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:57 vm10.local ceph-mon[53712]: pgmap v1736: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:23:59.146 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:23:59.146 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:23:59.173 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:23:59.173 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:23:59 vm11.local ceph-mon[53973]: pgmap v1737: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:23:59 vm10.local ceph-mon[53712]: pgmap v1737: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:01 vm11.local ceph-mon[53973]: pgmap v1738: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:01 vm10.local ceph-mon[53712]: pgmap v1738: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:03 vm11.local ceph-mon[53973]: pgmap v1739: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:04.175 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:04.175 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:03 vm10.local ceph-mon[53712]: pgmap v1739: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:04.202 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:04.203 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:05 vm10.local ceph-mon[53712]: pgmap v1740: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:05 vm11.local ceph-mon[53973]: pgmap v1740: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:07 vm11.local ceph-mon[53973]: pgmap v1741: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:07 vm10.local ceph-mon[53712]: pgmap v1741: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:09.204 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:09.205 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:09.233 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:09.233 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:09 vm11.local ceph-mon[53973]: pgmap v1742: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:09 vm10.local ceph-mon[53712]: pgmap v1742: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:11 vm11.local ceph-mon[53973]: pgmap v1743: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:11 vm10.local ceph-mon[53712]: pgmap v1743: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:13 vm11.local ceph-mon[53973]: pgmap v1744: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:13 vm10.local ceph-mon[53712]: pgmap v1744: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:14.235 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:14.235 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:14.289 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:14.289 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:15.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:15 vm10.local ceph-mon[53712]: pgmap v1745: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:15 vm11.local ceph-mon[53973]: pgmap v1745: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:17.925 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:17 vm11.local ceph-mon[53973]: pgmap v1746: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:18.117 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:17 vm10.local ceph-mon[53712]: pgmap v1746: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:24:19.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:24:19.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:24:19.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:24:19.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:24:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:24:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:24:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:24:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:24:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:24:19.290 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:19.291 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:19.316 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:19.317 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:19 vm11.local ceph-mon[53973]: pgmap v1747: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:19 vm10.local ceph-mon[53712]: pgmap v1747: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:21 vm11.local ceph-mon[53973]: pgmap v1748: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:21 vm10.local ceph-mon[53712]: pgmap v1748: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:23 vm10.local ceph-mon[53712]: pgmap v1749: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:24.318 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:24.319 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:23 vm11.local ceph-mon[53973]: pgmap v1749: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:24.344 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:24.345 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:25 vm10.local ceph-mon[53712]: pgmap v1750: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:25 vm11.local ceph-mon[53973]: pgmap v1750: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:27 vm11.local ceph-mon[53973]: pgmap v1751: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:27 vm10.local ceph-mon[53712]: pgmap v1751: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:24:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:24:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:24:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:24:29.346 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:29.347 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:29.372 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:29.372 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:29 vm11.local ceph-mon[53973]: pgmap v1752: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:29 vm10.local ceph-mon[53712]: pgmap v1752: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:31 vm11.local ceph-mon[53973]: pgmap v1753: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:31 vm10.local ceph-mon[53712]: pgmap v1753: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:33 vm11.local ceph-mon[53973]: pgmap v1754: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:33 vm10.local ceph-mon[53712]: pgmap v1754: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:34.374 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:34.374 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:34.401 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:34.401 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:35 vm10.local ceph-mon[53712]: pgmap v1755: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:35 vm11.local ceph-mon[53973]: pgmap v1755: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:37 vm11.local ceph-mon[53973]: pgmap v1756: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:37 vm10.local ceph-mon[53712]: pgmap v1756: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:39.402 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:39.403 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:39.429 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:39.430 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:39 vm11.local ceph-mon[53973]: pgmap v1757: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:39 vm10.local ceph-mon[53712]: pgmap v1757: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:41 vm11.local ceph-mon[53973]: pgmap v1758: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:41 vm10.local ceph-mon[53712]: pgmap v1758: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:43.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:42 vm10.local ceph-mon[53712]: pgmap v1759: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:42 vm11.local ceph-mon[53973]: pgmap v1759: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:44.431 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:44.431 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:44.457 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:44.457 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:45 vm10.local ceph-mon[53712]: pgmap v1760: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:45 vm11.local ceph-mon[53973]: pgmap v1760: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:47 vm11.local ceph-mon[53973]: pgmap v1761: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:47 vm10.local ceph-mon[53712]: pgmap v1761: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:49.459 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:49.459 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:49.485 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:49.485 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:49 vm11.local ceph-mon[53973]: pgmap v1762: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:49 vm10.local ceph-mon[53712]: pgmap v1762: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:51 vm11.local ceph-mon[53973]: pgmap v1763: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:51 vm10.local ceph-mon[53712]: pgmap v1763: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:53 vm11.local ceph-mon[53973]: pgmap v1764: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:53 vm10.local ceph-mon[53712]: pgmap v1764: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:54.487 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:54.487 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:54.513 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:54.514 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:24:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:55 vm10.local ceph-mon[53712]: pgmap v1765: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:55 vm11.local ceph-mon[53973]: pgmap v1765: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:24:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:57 vm11.local ceph-mon[53973]: pgmap v1766: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:57 vm10.local ceph-mon[53712]: pgmap v1766: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:24:59.516 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:24:59.516 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:24:59.544 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:24:59.544 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:24:59 vm11.local ceph-mon[53973]: pgmap v1767: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:24:59 vm10.local ceph-mon[53712]: pgmap v1767: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:02.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:01 vm11.local ceph-mon[53973]: pgmap v1768: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:01 vm10.local ceph-mon[53712]: pgmap v1768: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:03 vm11.local ceph-mon[53973]: pgmap v1769: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:03 vm10.local ceph-mon[53712]: pgmap v1769: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:04.546 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:04.546 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:04.573 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:04.574 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:05 vm10.local ceph-mon[53712]: pgmap v1770: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:05 vm11.local ceph-mon[53973]: pgmap v1770: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:07 vm11.local ceph-mon[53973]: pgmap v1771: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:08.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:07 vm10.local ceph-mon[53712]: pgmap v1771: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:09.575 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:09.576 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:09.603 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:09.604 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:10.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:09 vm10.local ceph-mon[53712]: pgmap v1772: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:09 vm11.local ceph-mon[53973]: pgmap v1772: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:11.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:10 vm10.local ceph-mon[53712]: pgmap v1773: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:10 vm11.local ceph-mon[53973]: pgmap v1773: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:13 vm11.local ceph-mon[53973]: pgmap v1774: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:13 vm10.local ceph-mon[53712]: pgmap v1774: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:14.605 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:14.606 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:14.633 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:14.634 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:15 vm10.local ceph-mon[53712]: pgmap v1775: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:15 vm11.local ceph-mon[53973]: pgmap v1775: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:17 vm11.local ceph-mon[53973]: pgmap v1776: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:18.177 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:17 vm10.local ceph-mon[53712]: pgmap v1776: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:25:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:25:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:25:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:25:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:25:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:25:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:25:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:25:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:25:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:25:19.635 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:19.636 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:19.661 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:19.662 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:19 vm11.local ceph-mon[53973]: pgmap v1777: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:19 vm10.local ceph-mon[53712]: pgmap v1777: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:21 vm11.local ceph-mon[53973]: pgmap v1778: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:21 vm10.local ceph-mon[53712]: pgmap v1778: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:23 vm11.local ceph-mon[53973]: pgmap v1779: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:23 vm10.local ceph-mon[53712]: pgmap v1779: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:24.663 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:24.664 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:24.694 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:24.695 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:25.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:25 vm10.local ceph-mon[53712]: pgmap v1780: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:25 vm11.local ceph-mon[53973]: pgmap v1780: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:27 vm11.local ceph-mon[53973]: pgmap v1781: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:27 vm10.local ceph-mon[53712]: pgmap v1781: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:25:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:25:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:25:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:25:29.696 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:29.696 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:29.722 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:29.723 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:29 vm11.local ceph-mon[53973]: pgmap v1782: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:29 vm10.local ceph-mon[53712]: pgmap v1782: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:31 vm11.local ceph-mon[53973]: pgmap v1783: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:31 vm10.local ceph-mon[53712]: pgmap v1783: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:33.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:32 vm10.local ceph-mon[53712]: pgmap v1784: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:32 vm11.local ceph-mon[53973]: pgmap v1784: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:34.724 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:34.725 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:34.754 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:34.754 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:35 vm10.local ceph-mon[53712]: pgmap v1785: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:35 vm11.local ceph-mon[53973]: pgmap v1785: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:37 vm11.local ceph-mon[53973]: pgmap v1786: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:37 vm10.local ceph-mon[53712]: pgmap v1786: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:39.756 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:39.756 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:39.783 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:39.784 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:39 vm11.local ceph-mon[53973]: pgmap v1787: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:39 vm10.local ceph-mon[53712]: pgmap v1787: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:41 vm11.local ceph-mon[53973]: pgmap v1788: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:41 vm10.local ceph-mon[53712]: pgmap v1788: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:43 vm11.local ceph-mon[53973]: pgmap v1789: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:43 vm10.local ceph-mon[53712]: pgmap v1789: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:44.785 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:44.786 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:44.813 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:44.814 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:45 vm10.local ceph-mon[53712]: pgmap v1790: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:45 vm11.local ceph-mon[53973]: pgmap v1790: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:47 vm11.local ceph-mon[53973]: pgmap v1791: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:47 vm10.local ceph-mon[53712]: pgmap v1791: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:49.815 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:49.816 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:49.846 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:49.846 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:49 vm11.local ceph-mon[53973]: pgmap v1792: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:49 vm10.local ceph-mon[53712]: pgmap v1792: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:51 vm11.local ceph-mon[53973]: pgmap v1793: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:51 vm10.local ceph-mon[53712]: pgmap v1793: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:53 vm11.local ceph-mon[53973]: pgmap v1794: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:53 vm10.local ceph-mon[53712]: pgmap v1794: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:54.848 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:54.848 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:54.875 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:54.876 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:25:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:55 vm10.local ceph-mon[53712]: pgmap v1795: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:55 vm11.local ceph-mon[53973]: pgmap v1795: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:25:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:57 vm11.local ceph-mon[53973]: pgmap v1796: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:57 vm10.local ceph-mon[53712]: pgmap v1796: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:25:59.877 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:25:59.878 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:25:59.906 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:25:59.907 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:25:59 vm11.local ceph-mon[53973]: pgmap v1797: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:25:59 vm10.local ceph-mon[53712]: pgmap v1797: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:01 vm11.local ceph-mon[53973]: pgmap v1798: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:01 vm10.local ceph-mon[53712]: pgmap v1798: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:03 vm11.local ceph-mon[53973]: pgmap v1799: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:03 vm10.local ceph-mon[53712]: pgmap v1799: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:04.908 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:04.909 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:04.934 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:04.934 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:05 vm10.local ceph-mon[53712]: pgmap v1800: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:05 vm11.local ceph-mon[53973]: pgmap v1800: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:07 vm11.local ceph-mon[53973]: pgmap v1801: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:07 vm10.local ceph-mon[53712]: pgmap v1801: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:09.936 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:09.936 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:09.962 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:09.963 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:09 vm11.local ceph-mon[53973]: pgmap v1802: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:09 vm10.local ceph-mon[53712]: pgmap v1802: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:11 vm11.local ceph-mon[53973]: pgmap v1803: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:11 vm10.local ceph-mon[53712]: pgmap v1803: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:13 vm11.local ceph-mon[53973]: pgmap v1804: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:13 vm10.local ceph-mon[53712]: pgmap v1804: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:14.964 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:14.965 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:14.990 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:14.991 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:15 vm11.local ceph-mon[53973]: pgmap v1805: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:15 vm10.local ceph-mon[53712]: pgmap v1805: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:17 vm11.local ceph-mon[53973]: pgmap v1806: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:18.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:17 vm10.local ceph-mon[53712]: pgmap v1806: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:18.781 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:26:18.781 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:26:18.781 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:26:18.781 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:26:18.781 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:26:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:26:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:26:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:26:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:26:19.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:26:19.992 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:19.993 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:20.019 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:20.020 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:19 vm11.local ceph-mon[53973]: pgmap v1807: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:19 vm10.local ceph-mon[53712]: pgmap v1807: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:21 vm11.local ceph-mon[53973]: pgmap v1808: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:21 vm10.local ceph-mon[53712]: pgmap v1808: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:23 vm11.local ceph-mon[53973]: pgmap v1809: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:23 vm10.local ceph-mon[53712]: pgmap v1809: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:25.021 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:25.022 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:25.049 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:25.050 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:25 vm10.local ceph-mon[53712]: pgmap v1810: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:25 vm11.local ceph-mon[53973]: pgmap v1810: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:27 vm11.local ceph-mon[53973]: pgmap v1811: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:27 vm10.local ceph-mon[53712]: pgmap v1811: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:26:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:26:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:26:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:26:30.051 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:30.052 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:30.247 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:30.248 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:30.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:30 vm10.local ceph-mon[53712]: pgmap v1812: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:30.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:30 vm11.local ceph-mon[53973]: pgmap v1812: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:31 vm10.local ceph-mon[53712]: pgmap v1813: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:31.492 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:31 vm11.local ceph-mon[53973]: pgmap v1813: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:33 vm11.local ceph-mon[53973]: pgmap v1814: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:33 vm10.local ceph-mon[53712]: pgmap v1814: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:35.249 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:35.250 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:35.276 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:35.276 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:35 vm11.local ceph-mon[53973]: pgmap v1815: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:35 vm10.local ceph-mon[53712]: pgmap v1815: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:37 vm11.local ceph-mon[53973]: pgmap v1816: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:37 vm10.local ceph-mon[53712]: pgmap v1816: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:39 vm11.local ceph-mon[53973]: pgmap v1817: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:39 vm10.local ceph-mon[53712]: pgmap v1817: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:40.278 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:40.278 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:40.304 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:40.305 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:41 vm10.local ceph-mon[53712]: pgmap v1818: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:41 vm11.local ceph-mon[53973]: pgmap v1818: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:42 vm10.local ceph-mon[53712]: pgmap v1819: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:42 vm11.local ceph-mon[53973]: pgmap v1819: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:45.306 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:45.307 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:45.334 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:45.335 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:45 vm11.local ceph-mon[53973]: pgmap v1820: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:45 vm10.local ceph-mon[53712]: pgmap v1820: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:47 vm11.local ceph-mon[53973]: pgmap v1821: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:47 vm10.local ceph-mon[53712]: pgmap v1821: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:49 vm11.local ceph-mon[53973]: pgmap v1822: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:49 vm10.local ceph-mon[53712]: pgmap v1822: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:50.337 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:50.337 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:50.365 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:50.366 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:51 vm11.local ceph-mon[53973]: pgmap v1823: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:51 vm10.local ceph-mon[53712]: pgmap v1823: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:53 vm11.local ceph-mon[53973]: pgmap v1824: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:53 vm10.local ceph-mon[53712]: pgmap v1824: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:55.367 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:26:55.367 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:26:55.393 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:26:55.393 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:26:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:55 vm11.local ceph-mon[53973]: pgmap v1825: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:55 vm10.local ceph-mon[53712]: pgmap v1825: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:26:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:57 vm11.local ceph-mon[53973]: pgmap v1826: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:26:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:57 vm10.local ceph-mon[53712]: pgmap v1826: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:26:59 vm11.local ceph-mon[53973]: pgmap v1827: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:00.187 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:26:59 vm10.local ceph-mon[53712]: pgmap v1827: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:00.394 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:00.395 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:00.582 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:00.582 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:01 vm11.local ceph-mon[53973]: pgmap v1828: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:01 vm10.local ceph-mon[53712]: pgmap v1828: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:03 vm11.local ceph-mon[53973]: pgmap v1829: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:03 vm10.local ceph-mon[53712]: pgmap v1829: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:05.584 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:05.584 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:05 vm11.local ceph-mon[53973]: pgmap v1830: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:05.611 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:05.611 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:05 vm10.local ceph-mon[53712]: pgmap v1830: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:07 vm11.local ceph-mon[53973]: pgmap v1831: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:07 vm10.local ceph-mon[53712]: pgmap v1831: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:09 vm11.local ceph-mon[53973]: pgmap v1832: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:09 vm10.local ceph-mon[53712]: pgmap v1832: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:10.613 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:10.613 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:10.645 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:10.646 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:11 vm11.local ceph-mon[53973]: pgmap v1833: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:11 vm10.local ceph-mon[53712]: pgmap v1833: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:13 vm11.local ceph-mon[53973]: pgmap v1834: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:13 vm10.local ceph-mon[53712]: pgmap v1834: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:15 vm11.local ceph-mon[53973]: pgmap v1835: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:15.647 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:15.648 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:15.675 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:15.675 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:15 vm10.local ceph-mon[53712]: pgmap v1835: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:17 vm11.local ceph-mon[53973]: pgmap v1836: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:17 vm10.local ceph-mon[53712]: pgmap v1836: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:18.976 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:27:18.977 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:27:18.977 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:18 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:27:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:27:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:27:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:18 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:27:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:19 vm11.local ceph-mon[53973]: pgmap v1837: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:19 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:27:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:19 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:27:20.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:19 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:27:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:19 vm10.local ceph-mon[53712]: pgmap v1837: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:19 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:27:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:19 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:27:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:19 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:27:20.677 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:20.677 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:20.719 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:20.719 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:22.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:21 vm11.local ceph-mon[53973]: pgmap v1838: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:21 vm10.local ceph-mon[53712]: pgmap v1838: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:24.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:23 vm11.local ceph-mon[53973]: pgmap v1839: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:23 vm10.local ceph-mon[53712]: pgmap v1839: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:25 vm11.local ceph-mon[53973]: pgmap v1840: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:25.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:25 vm10.local ceph-mon[53712]: pgmap v1840: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:25.721 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:25.721 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:25.748 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:25.749 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:27 vm11.local ceph-mon[53973]: pgmap v1841: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:27 vm10.local ceph-mon[53712]: pgmap v1841: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:27:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:27:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:27:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:27:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:29 vm10.local ceph-mon[53712]: pgmap v1842: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:29 vm11.local ceph-mon[53973]: pgmap v1842: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:30.751 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:30.751 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:30.779 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:30.780 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:31.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:30 vm10.local ceph-mon[53712]: pgmap v1843: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:31.242 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:30 vm11.local ceph-mon[53973]: pgmap v1843: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:33 vm11.local ceph-mon[53973]: pgmap v1844: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:33 vm10.local ceph-mon[53712]: pgmap v1844: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:35 vm11.local ceph-mon[53973]: pgmap v1845: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:35 vm10.local ceph-mon[53712]: pgmap v1845: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:35.781 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:35.782 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:35.809 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:35.809 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:36 vm11.local ceph-mon[53973]: pgmap v1846: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:36 vm10.local ceph-mon[53712]: pgmap v1846: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:39 vm11.local ceph-mon[53973]: pgmap v1847: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:39 vm10.local ceph-mon[53712]: pgmap v1847: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:40.811 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:40.811 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:40.837 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:40.837 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:41 vm11.local ceph-mon[53973]: pgmap v1848: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:41 vm10.local ceph-mon[53712]: pgmap v1848: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:43 vm11.local ceph-mon[53973]: pgmap v1849: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:43 vm10.local ceph-mon[53712]: pgmap v1849: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:45 vm11.local ceph-mon[53973]: pgmap v1850: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:45 vm10.local ceph-mon[53712]: pgmap v1850: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:45.839 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:45.840 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:45.867 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:45.867 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:47 vm11.local ceph-mon[53973]: pgmap v1851: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:47 vm10.local ceph-mon[53712]: pgmap v1851: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:49 vm11.local ceph-mon[53973]: pgmap v1852: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:49 vm10.local ceph-mon[53712]: pgmap v1852: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:50.869 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:50.869 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:50.896 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:50.896 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:51 vm11.local ceph-mon[53973]: pgmap v1853: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:51 vm10.local ceph-mon[53712]: pgmap v1853: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:54.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:53 vm11.local ceph-mon[53973]: pgmap v1854: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:53 vm10.local ceph-mon[53712]: pgmap v1854: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:55 vm11.local ceph-mon[53973]: pgmap v1855: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:55 vm10.local ceph-mon[53712]: pgmap v1855: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:27:55.898 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:27:55.898 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:27:55.926 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:27:55.926 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:27:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:57 vm11.local ceph-mon[53973]: pgmap v1856: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:27:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:57 vm10.local ceph-mon[53712]: pgmap v1856: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:27:59 vm11.local ceph-mon[53973]: pgmap v1857: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:00.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:27:59 vm10.local ceph-mon[53712]: pgmap v1857: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:00.928 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:00.928 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:00.956 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:00.956 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:01 vm11.local ceph-mon[53973]: pgmap v1858: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:01 vm10.local ceph-mon[53712]: pgmap v1858: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:03 vm11.local ceph-mon[53973]: pgmap v1859: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:03 vm10.local ceph-mon[53712]: pgmap v1859: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:05 vm11.local ceph-mon[53973]: pgmap v1860: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:05 vm10.local ceph-mon[53712]: pgmap v1860: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:05.957 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:05.958 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:05.986 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:05.987 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:07 vm11.local ceph-mon[53973]: pgmap v1861: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:07 vm10.local ceph-mon[53712]: pgmap v1861: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:09 vm11.local ceph-mon[53973]: pgmap v1862: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:09 vm10.local ceph-mon[53712]: pgmap v1862: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:10.988 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:10.989 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:11.015 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:11.016 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:11 vm11.local ceph-mon[53973]: pgmap v1863: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:11 vm10.local ceph-mon[53712]: pgmap v1863: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:13 vm11.local ceph-mon[53973]: pgmap v1864: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:13 vm10.local ceph-mon[53712]: pgmap v1864: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:15 vm11.local ceph-mon[53973]: pgmap v1865: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:15 vm10.local ceph-mon[53712]: pgmap v1865: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:16.018 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:16.018 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:16.046 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:16.046 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:17 vm11.local ceph-mon[53973]: pgmap v1866: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:17 vm10.local ceph-mon[53712]: pgmap v1866: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:19 vm11.local ceph-mon[53973]: pgmap v1867: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:19 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:28:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:19 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:28:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:19 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:28:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:19 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:19 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:19 vm10.local ceph-mon[53712]: pgmap v1867: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:20.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:19 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:28:20.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:19 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:28:20.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:19 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:28:20.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:19 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:20.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:19 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:21.048 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:21.048 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:21.080 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:21.081 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:20 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:21.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:20 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:21.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:20 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:21.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:20 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:21.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:20 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:21.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:20 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:28:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:21 vm10.local ceph-mon[53712]: pgmap v1868: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:21 vm11.local ceph-mon[53973]: pgmap v1868: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:22 vm10.local ceph-mon[53712]: pgmap v1869: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:22 vm11.local ceph-mon[53973]: pgmap v1869: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:25 vm10.local ceph-mon[53712]: pgmap v1870: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:25 vm11.local ceph-mon[53973]: pgmap v1870: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:26.082 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:26.083 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:26.111 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:26.112 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:27 vm11.local ceph-mon[53973]: pgmap v1871: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:27 vm10.local ceph-mon[53712]: pgmap v1871: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:28:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:28:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:28:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:28:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:29 vm11.local ceph-mon[53973]: pgmap v1872: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:29 vm10.local ceph-mon[53712]: pgmap v1872: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:31.113 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:31.114 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:31.142 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:31.142 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:31 vm11.local ceph-mon[53973]: pgmap v1873: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:31 vm10.local ceph-mon[53712]: pgmap v1873: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:33 vm11.local ceph-mon[53973]: pgmap v1874: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:33 vm10.local ceph-mon[53712]: pgmap v1874: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:35 vm10.local ceph-mon[53712]: pgmap v1875: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:35 vm11.local ceph-mon[53973]: pgmap v1875: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:36.144 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:36.144 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:36.172 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:36.172 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:37 vm11.local ceph-mon[53973]: pgmap v1876: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:37 vm10.local ceph-mon[53712]: pgmap v1876: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:39 vm11.local ceph-mon[53973]: pgmap v1877: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:39 vm10.local ceph-mon[53712]: pgmap v1877: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:41.173 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:41.174 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:41.203 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:41.204 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:41 vm11.local ceph-mon[53973]: pgmap v1878: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:41 vm10.local ceph-mon[53712]: pgmap v1878: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:43 vm11.local ceph-mon[53973]: pgmap v1879: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:44.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:43 vm10.local ceph-mon[53712]: pgmap v1879: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:45.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:45 vm10.local ceph-mon[53712]: pgmap v1880: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:45 vm11.local ceph-mon[53973]: pgmap v1880: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:46.205 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:46.206 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:46.234 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:46.235 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:47 vm11.local ceph-mon[53973]: pgmap v1881: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:47 vm10.local ceph-mon[53712]: pgmap v1881: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:49 vm11.local ceph-mon[53973]: pgmap v1882: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:49 vm10.local ceph-mon[53712]: pgmap v1882: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:51.236 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:51.237 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:51.266 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:51.266 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:51 vm11.local ceph-mon[53973]: pgmap v1883: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:51 vm10.local ceph-mon[53712]: pgmap v1883: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:54.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:53 vm11.local ceph-mon[53973]: pgmap v1884: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:53 vm10.local ceph-mon[53712]: pgmap v1884: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:55 vm10.local ceph-mon[53712]: pgmap v1885: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:55 vm11.local ceph-mon[53973]: pgmap v1885: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:28:56.268 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:28:56.268 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:28:56.295 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:28:56.296 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:28:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:57 vm11.local ceph-mon[53973]: pgmap v1886: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:28:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:57 vm10.local ceph-mon[53712]: pgmap v1886: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:28:59 vm11.local ceph-mon[53973]: pgmap v1887: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:28:59 vm10.local ceph-mon[53712]: pgmap v1887: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:01.298 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:01.298 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:01.326 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:01.327 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:02.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:01 vm11.local ceph-mon[53973]: pgmap v1888: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:01 vm10.local ceph-mon[53712]: pgmap v1888: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:04.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:03 vm11.local ceph-mon[53973]: pgmap v1889: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:04.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:03 vm10.local ceph-mon[53712]: pgmap v1889: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:05 vm10.local ceph-mon[53712]: pgmap v1890: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:05 vm11.local ceph-mon[53973]: pgmap v1890: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:06.328 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:06.328 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:06.354 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:06.354 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:07 vm11.local ceph-mon[53973]: pgmap v1891: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:07 vm10.local ceph-mon[53712]: pgmap v1891: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:09 vm11.local ceph-mon[53973]: pgmap v1892: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:09 vm10.local ceph-mon[53712]: pgmap v1892: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:11.355 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:11.356 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:11.389 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:11.389 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:11 vm11.local ceph-mon[53973]: pgmap v1893: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:11 vm10.local ceph-mon[53712]: pgmap v1893: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:14.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:13 vm11.local ceph-mon[53973]: pgmap v1894: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:13 vm10.local ceph-mon[53712]: pgmap v1894: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:15 vm10.local ceph-mon[53712]: pgmap v1895: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:15 vm11.local ceph-mon[53973]: pgmap v1895: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:16.390 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:16.391 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:16.417 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:16.418 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:17 vm11.local ceph-mon[53973]: pgmap v1896: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:17 vm10.local ceph-mon[53712]: pgmap v1896: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:19 vm11.local ceph-mon[53973]: pgmap v1897: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:20.122 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:19 vm10.local ceph-mon[53712]: pgmap v1897: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:20 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:29:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:20 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:29:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:20 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:29:21.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:20 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:29:21.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:20 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:29:21.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:20 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:29:21.422 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:21.422 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:21.452 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:21.452 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:22.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:21 vm10.local ceph-mon[53712]: pgmap v1898: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:22.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:21 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:29:22.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:21 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:29:22.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:21 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:29:22.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:21 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:29:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:21 vm11.local ceph-mon[53973]: pgmap v1898: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:21 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:29:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:21 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:29:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:21 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:29:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:21 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:29:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:23 vm10.local ceph-mon[53712]: pgmap v1899: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:23 vm11.local ceph-mon[53973]: pgmap v1899: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:25 vm10.local ceph-mon[53712]: pgmap v1900: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:25 vm11.local ceph-mon[53973]: pgmap v1900: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:26.454 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:26.455 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:26.483 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:26.484 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:27 vm11.local ceph-mon[53973]: pgmap v1901: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:27 vm10.local ceph-mon[53712]: pgmap v1901: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:29:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:29:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:29:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:29:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:29 vm11.local ceph-mon[53973]: pgmap v1902: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:29:30.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:29 vm10.local ceph-mon[53712]: pgmap v1902: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:29:31.485 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:31.486 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:31.512 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:31.513 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:31 vm11.local ceph-mon[53973]: pgmap v1903: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:29:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:31 vm10.local ceph-mon[53712]: pgmap v1903: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:29:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:33 vm10.local ceph-mon[53712]: pgmap v1904: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:33 vm11.local ceph-mon[53973]: pgmap v1904: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:35 vm10.local ceph-mon[53712]: pgmap v1905: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:29:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:35 vm11.local ceph-mon[53973]: pgmap v1905: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:29:36.515 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:36.515 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:36.618 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:36.618 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:36 vm11.local ceph-mon[53973]: pgmap v1906: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:36 vm10.local ceph-mon[53712]: pgmap v1906: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:39 vm11.local ceph-mon[53973]: pgmap v1907: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:40.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:39 vm10.local ceph-mon[53712]: pgmap v1907: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:41.620 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:41.620 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:41.648 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:41.649 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:42.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:41 vm11.local ceph-mon[53973]: pgmap v1908: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:41 vm10.local ceph-mon[53712]: pgmap v1908: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:43 vm11.local ceph-mon[53973]: pgmap v1909: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:43 vm10.local ceph-mon[53712]: pgmap v1909: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:45 vm10.local ceph-mon[53712]: pgmap v1910: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:45 vm11.local ceph-mon[53973]: pgmap v1910: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:46.651 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:46.651 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:46.679 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:46.679 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:47 vm11.local ceph-mon[53973]: pgmap v1911: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:47 vm10.local ceph-mon[53712]: pgmap v1911: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:50.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:49 vm11.local ceph-mon[53973]: pgmap v1912: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:49 vm10.local ceph-mon[53712]: pgmap v1912: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:51.681 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:51.682 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:51.708 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:51.709 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:52 vm11.local ceph-mon[53973]: pgmap v1913: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:52.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:51 vm10.local ceph-mon[53712]: pgmap v1913: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:52 vm11.local ceph-mon[53973]: pgmap v1914: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:52 vm10.local ceph-mon[53712]: pgmap v1914: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:55 vm10.local ceph-mon[53712]: pgmap v1915: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:55 vm11.local ceph-mon[53973]: pgmap v1915: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:29:56.710 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:29:56.711 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:29:56.739 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:29:56.740 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:29:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:57 vm11.local ceph-mon[53973]: pgmap v1916: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:29:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:57 vm10.local ceph-mon[53712]: pgmap v1916: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:00.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:29:59 vm11.local ceph-mon[53973]: pgmap v1917: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:29:59 vm10.local ceph-mon[53712]: pgmap v1917: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:00 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T18:30:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:00 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T18:30:01.742 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:01.743 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:01.770 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:01.771 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:01 vm10.local ceph-mon[53712]: pgmap v1918: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:01 vm11.local ceph-mon[53973]: pgmap v1918: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:03 vm10.local ceph-mon[53712]: pgmap v1919: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:04.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:03 vm11.local ceph-mon[53973]: pgmap v1919: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:05 vm10.local ceph-mon[53712]: pgmap v1920: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:05 vm11.local ceph-mon[53973]: pgmap v1920: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:06.772 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:06.772 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:06.800 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:06.801 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:07 vm11.local ceph-mon[53973]: pgmap v1921: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:07 vm10.local ceph-mon[53712]: pgmap v1921: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:10.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:09 vm11.local ceph-mon[53973]: pgmap v1922: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:09 vm10.local ceph-mon[53712]: pgmap v1922: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:11.802 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:11.803 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:11.828 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:11.829 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:12.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:11 vm11.local ceph-mon[53973]: pgmap v1923: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:11 vm10.local ceph-mon[53712]: pgmap v1923: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:13 vm10.local ceph-mon[53712]: pgmap v1924: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:13 vm11.local ceph-mon[53973]: pgmap v1924: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:15 vm10.local ceph-mon[53712]: pgmap v1925: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:15 vm11.local ceph-mon[53973]: pgmap v1925: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:16.830 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:16.830 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:16.857 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:16.857 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:17 vm11.local ceph-mon[53973]: pgmap v1926: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:17 vm10.local ceph-mon[53712]: pgmap v1926: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:20.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:19 vm11.local ceph-mon[53973]: pgmap v1927: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:19 vm10.local ceph-mon[53712]: pgmap v1927: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:21.859 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:21.859 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:21.895 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:21.896 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:22.084 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:21 vm10.local ceph-mon[53712]: pgmap v1928: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:22.084 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:21 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:30:22.084 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:21 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:30:22.084 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:21 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:30:22.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:21 vm11.local ceph-mon[53973]: pgmap v1928: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:22.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:21 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:30:22.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:21 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:30:22.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:21 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:30:23.836 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:23 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:23.836 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:23 vm10.local ceph-mon[53712]: pgmap v1929: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:23.836 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:23 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:23 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:23 vm11.local ceph-mon[53973]: pgmap v1929: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:23.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:23 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:30:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:24 vm11.local ceph-mon[53973]: pgmap v1930: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:30:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:30:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:24 vm10.local ceph-mon[53712]: pgmap v1930: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:26.897 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:26.898 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:26.927 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:26.928 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:27 vm11.local ceph-mon[53973]: pgmap v1931: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:27 vm10.local ceph-mon[53712]: pgmap v1931: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:30:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:30:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:30:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:30:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:29 vm10.local ceph-mon[53712]: pgmap v1932: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:29 vm11.local ceph-mon[53973]: pgmap v1932: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:31.929 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:31.930 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:31.956 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:31.956 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:32.142 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:31 vm11.local ceph-mon[53973]: pgmap v1933: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:31 vm10.local ceph-mon[53712]: pgmap v1933: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:33 vm10.local ceph-mon[53712]: pgmap v1934: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:33 vm11.local ceph-mon[53973]: pgmap v1934: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:35 vm10.local ceph-mon[53712]: pgmap v1935: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:35 vm11.local ceph-mon[53973]: pgmap v1935: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:36.958 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:36.958 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:36.984 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:36.985 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:37 vm11.local ceph-mon[53973]: pgmap v1936: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:37 vm10.local ceph-mon[53712]: pgmap v1936: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:39 vm11.local ceph-mon[53973]: pgmap v1937: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:39 vm10.local ceph-mon[53712]: pgmap v1937: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:41.986 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:41.987 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:42.013 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:42.013 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:42.171 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:41 vm11.local ceph-mon[53973]: pgmap v1938: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:41 vm10.local ceph-mon[53712]: pgmap v1938: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:43 vm10.local ceph-mon[53712]: pgmap v1939: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:43 vm11.local ceph-mon[53973]: pgmap v1939: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:45 vm10.local ceph-mon[53712]: pgmap v1940: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:45 vm11.local ceph-mon[53973]: pgmap v1940: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:47.014 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:47.015 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:47.040 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:47.041 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:47 vm11.local ceph-mon[53973]: pgmap v1941: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:47 vm10.local ceph-mon[53712]: pgmap v1941: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:49 vm10.local ceph-mon[53712]: pgmap v1942: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:49 vm11.local ceph-mon[53973]: pgmap v1942: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:52.042 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:52.042 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:52.068 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:52.069 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:51 vm10.local ceph-mon[53712]: pgmap v1943: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:51 vm11.local ceph-mon[53973]: pgmap v1943: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:53 vm10.local ceph-mon[53712]: pgmap v1944: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:53 vm11.local ceph-mon[53973]: pgmap v1944: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:55 vm10.local ceph-mon[53712]: pgmap v1945: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:55 vm11.local ceph-mon[53973]: pgmap v1945: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:30:57.070 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:30:57.071 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:30:57.096 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:30:57.096 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:30:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:57 vm11.local ceph-mon[53973]: pgmap v1946: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:30:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:57 vm10.local ceph-mon[53712]: pgmap v1946: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:00.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:30:59 vm10.local ceph-mon[53712]: pgmap v1947: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:30:59 vm11.local ceph-mon[53973]: pgmap v1947: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:02.098 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:02.099 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:02.125 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:02.125 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:02.143 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:01 vm11.local ceph-mon[53973]: pgmap v1948: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:01 vm10.local ceph-mon[53712]: pgmap v1948: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:03 vm10.local ceph-mon[53712]: pgmap v1949: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:04.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:03 vm11.local ceph-mon[53973]: pgmap v1949: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:05 vm10.local ceph-mon[53712]: pgmap v1950: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:05 vm11.local ceph-mon[53973]: pgmap v1950: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:07.127 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:07.127 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:07.153 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:07.153 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:07 vm11.local ceph-mon[53973]: pgmap v1951: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:07 vm10.local ceph-mon[53712]: pgmap v1951: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:09 vm10.local ceph-mon[53712]: pgmap v1952: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:09 vm11.local ceph-mon[53973]: pgmap v1952: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:12.155 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:12.155 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:12.181 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:12.182 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:11 vm10.local ceph-mon[53712]: pgmap v1953: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:11 vm11.local ceph-mon[53973]: pgmap v1953: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:13 vm10.local ceph-mon[53712]: pgmap v1954: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:13 vm11.local ceph-mon[53973]: pgmap v1954: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:15 vm10.local ceph-mon[53712]: pgmap v1955: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:15 vm11.local ceph-mon[53973]: pgmap v1955: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:17.183 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:17.183 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:17.209 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:17.210 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:17 vm11.local ceph-mon[53973]: pgmap v1956: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:17 vm10.local ceph-mon[53712]: pgmap v1956: 97 pgs: 97 active+clean; 453 KiB data, 68 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:19 vm10.local ceph-mon[53712]: pgmap v1957: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:19 vm11.local ceph-mon[53973]: pgmap v1957: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:21 vm10.local ceph-mon[53712]: pgmap v1958: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:22.211 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:22.212 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:22.238 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:22.238 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:21 vm11.local ceph-mon[53973]: pgmap v1958: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:24.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:23 vm10.local ceph-mon[53712]: pgmap v1959: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:24.208 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:23 vm11.local ceph-mon[53973]: pgmap v1959: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:31:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:31:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:31:25.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:31:25.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:31:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:31:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:31:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:31:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:31:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:31:26.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:25 vm10.local ceph-mon[53712]: pgmap v1960: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:26.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:25 vm11.local ceph-mon[53973]: pgmap v1960: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:27.240 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:27.240 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:27.270 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:27.270 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:27 vm11.local ceph-mon[53973]: pgmap v1961: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:28.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:27 vm10.local ceph-mon[53712]: pgmap v1961: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:31:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:31:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:31:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:31:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:29 vm10.local ceph-mon[53712]: pgmap v1962: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:29 vm11.local ceph-mon[53973]: pgmap v1962: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:32.142 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:31 vm11.local ceph-mon[53973]: pgmap v1963: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:31 vm10.local ceph-mon[53712]: pgmap v1963: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:32.271 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:32.272 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:32.300 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:32.300 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:32 vm11.local ceph-mon[53973]: pgmap v1964: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:32 vm10.local ceph-mon[53712]: pgmap v1964: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:35 vm10.local ceph-mon[53712]: pgmap v1965: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:35 vm11.local ceph-mon[53973]: pgmap v1965: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:37.301 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:37.302 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:37.327 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:37.328 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:37 vm11.local ceph-mon[53973]: pgmap v1966: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:37 vm10.local ceph-mon[53712]: pgmap v1966: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:39 vm10.local ceph-mon[53712]: pgmap v1967: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:39 vm11.local ceph-mon[53973]: pgmap v1967: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:41 vm10.local ceph-mon[53712]: pgmap v1968: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:42.329 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:42.329 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:41 vm11.local ceph-mon[53973]: pgmap v1968: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:42.356 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:42.357 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:43 vm10.local ceph-mon[53712]: pgmap v1969: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:43 vm11.local ceph-mon[53973]: pgmap v1969: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:45 vm10.local ceph-mon[53712]: pgmap v1970: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:45 vm11.local ceph-mon[53973]: pgmap v1970: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:47.358 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:47.358 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:47.385 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:47.385 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:47 vm11.local ceph-mon[53973]: pgmap v1971: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:47 vm10.local ceph-mon[53712]: pgmap v1971: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:50.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:49 vm10.local ceph-mon[53712]: pgmap v1972: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:49 vm11.local ceph-mon[53973]: pgmap v1972: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:51 vm10.local ceph-mon[53712]: pgmap v1973: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:51 vm11.local ceph-mon[53973]: pgmap v1973: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:52.387 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:52.387 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:52.416 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:52.416 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:53 vm10.local ceph-mon[53712]: pgmap v1974: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:53 vm11.local ceph-mon[53973]: pgmap v1974: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:55 vm10.local ceph-mon[53712]: pgmap v1975: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:55 vm11.local ceph-mon[53973]: pgmap v1975: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:31:57.418 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:31:57.418 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:31:57.470 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:31:57.471 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:31:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:57 vm11.local ceph-mon[53973]: pgmap v1976: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:31:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:57 vm10.local ceph-mon[53712]: pgmap v1976: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:31:59 vm10.local ceph-mon[53712]: pgmap v1977: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:31:59 vm11.local ceph-mon[53973]: pgmap v1977: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:00 vm11.local ceph-mon[53973]: pgmap v1978: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:00 vm10.local ceph-mon[53712]: pgmap v1978: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:02.472 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:02.473 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:02.500 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:02.500 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:02 vm11.local ceph-mon[53973]: pgmap v1979: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:02 vm10.local ceph-mon[53712]: pgmap v1979: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:05 vm10.local ceph-mon[53712]: pgmap v1980: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:05 vm11.local ceph-mon[53973]: pgmap v1980: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:07.502 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:07.502 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:07.527 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:07.528 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:07 vm11.local ceph-mon[53973]: pgmap v1981: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:07 vm10.local ceph-mon[53712]: pgmap v1981: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:09 vm10.local ceph-mon[53712]: pgmap v1982: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:09 vm11.local ceph-mon[53973]: pgmap v1982: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:11 vm10.local ceph-mon[53712]: pgmap v1983: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:11 vm11.local ceph-mon[53973]: pgmap v1983: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:12.529 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:12.530 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:12.642 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:12.643 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:12 vm11.local ceph-mon[53973]: pgmap v1984: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:12 vm10.local ceph-mon[53712]: pgmap v1984: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:15 vm10.local ceph-mon[53712]: pgmap v1985: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:15 vm11.local ceph-mon[53973]: pgmap v1985: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:17.644 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:17.644 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:17.669 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:17.670 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:17 vm11.local ceph-mon[53973]: pgmap v1986: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:17 vm10.local ceph-mon[53712]: pgmap v1986: 97 pgs: 97 active+clean; 453 KiB data, 72 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:19 vm11.local ceph-mon[53973]: pgmap v1987: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:19.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:19 vm10.local ceph-mon[53712]: pgmap v1987: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:21 vm10.local ceph-mon[53712]: pgmap v1988: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:21 vm11.local ceph-mon[53973]: pgmap v1988: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:22.671 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:22.671 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:22.703 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:22.704 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:23 vm10.local ceph-mon[53712]: pgmap v1989: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:24.267 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:23 vm11.local ceph-mon[53973]: pgmap v1989: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:32:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:32:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:32:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:32:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:24 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:32:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:32:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:32:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:32:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:32:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:24 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:32:26.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:25 vm10.local ceph-mon[53712]: pgmap v1990: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:26.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:25 vm11.local ceph-mon[53973]: pgmap v1990: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:27.705 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:27.706 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:27.735 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:27.735 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:27 vm11.local ceph-mon[53973]: pgmap v1991: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:27 vm10.local ceph-mon[53712]: pgmap v1991: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:32:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:32:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:32:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:32:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:29 vm10.local ceph-mon[53712]: pgmap v1992: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:29 vm11.local ceph-mon[53973]: pgmap v1992: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:31 vm10.local ceph-mon[53712]: pgmap v1993: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:31 vm11.local ceph-mon[53973]: pgmap v1993: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:32.737 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:32.737 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:32.766 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:32.767 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:33 vm10.local ceph-mon[53712]: pgmap v1994: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:33 vm11.local ceph-mon[53973]: pgmap v1994: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:35 vm10.local ceph-mon[53712]: pgmap v1995: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:35 vm11.local ceph-mon[53973]: pgmap v1995: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:37.768 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:37.769 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:37.796 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:37.796 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:37 vm11.local ceph-mon[53973]: pgmap v1996: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:37 vm10.local ceph-mon[53712]: pgmap v1996: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:39 vm10.local ceph-mon[53712]: pgmap v1997: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:39 vm11.local ceph-mon[53973]: pgmap v1997: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:41 vm10.local ceph-mon[53712]: pgmap v1998: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:41 vm11.local ceph-mon[53973]: pgmap v1998: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:42.797 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:42.798 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:42.859 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:42.860 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:43 vm10.local ceph-mon[53712]: pgmap v1999: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:43 vm11.local ceph-mon[53973]: pgmap v1999: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:45 vm10.local ceph-mon[53712]: pgmap v2000: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:45 vm11.local ceph-mon[53973]: pgmap v2000: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:47.861 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:47.861 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:47.968 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:47.968 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:47 vm11.local ceph-mon[53973]: pgmap v2001: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:47 vm10.local ceph-mon[53712]: pgmap v2001: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:49 vm10.local ceph-mon[53712]: pgmap v2002: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:49 vm11.local ceph-mon[53973]: pgmap v2002: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:51 vm10.local ceph-mon[53712]: pgmap v2003: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:51 vm11.local ceph-mon[53973]: pgmap v2003: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:52.970 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:52.970 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:52.996 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:52.996 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:53 vm10.local ceph-mon[53712]: pgmap v2004: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:53 vm11.local ceph-mon[53973]: pgmap v2004: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:55 vm10.local ceph-mon[53712]: pgmap v2005: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:55 vm11.local ceph-mon[53973]: pgmap v2005: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:32:57.998 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:32:57.998 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:32:58.024 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:32:58.024 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:32:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:57 vm11.local ceph-mon[53973]: pgmap v2006: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:32:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:57 vm10.local ceph-mon[53712]: pgmap v2006: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:32:59 vm10.local ceph-mon[53712]: pgmap v2007: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:32:59 vm11.local ceph-mon[53973]: pgmap v2007: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:01 vm10.local ceph-mon[53712]: pgmap v2008: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:01 vm11.local ceph-mon[53973]: pgmap v2008: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:03.025 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:03.026 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:03.054 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:03.055 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:04.142 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:03 vm11.local ceph-mon[53973]: pgmap v2009: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:03 vm10.local ceph-mon[53712]: pgmap v2009: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:05 vm10.local ceph-mon[53712]: pgmap v2010: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:05 vm11.local ceph-mon[53973]: pgmap v2010: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:08.056 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:08.057 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:08.082 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:08.083 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:07 vm11.local ceph-mon[53973]: pgmap v2011: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:07 vm10.local ceph-mon[53712]: pgmap v2011: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:09 vm10.local ceph-mon[53712]: pgmap v2012: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:09 vm11.local ceph-mon[53973]: pgmap v2012: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:11 vm10.local ceph-mon[53712]: pgmap v2013: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:11 vm11.local ceph-mon[53973]: pgmap v2013: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:13.084 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:13.084 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:13.111 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:13.111 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:13 vm10.local ceph-mon[53712]: pgmap v2014: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:13 vm11.local ceph-mon[53973]: pgmap v2014: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:15 vm11.local ceph-mon[53973]: pgmap v2015: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:15 vm10.local ceph-mon[53712]: pgmap v2015: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:17 vm11.local ceph-mon[53973]: pgmap v2016: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:18.112 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:18.113 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:18.140 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:18.140 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:17 vm10.local ceph-mon[53712]: pgmap v2016: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:19 vm10.local ceph-mon[53712]: pgmap v2017: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:19 vm11.local ceph-mon[53973]: pgmap v2017: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:21 vm10.local ceph-mon[53712]: pgmap v2018: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:21 vm11.local ceph-mon[53973]: pgmap v2018: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:23.141 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:23.142 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:23.168 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:23.168 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:23 vm10.local ceph-mon[53712]: pgmap v2019: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:23 vm11.local ceph-mon[53973]: pgmap v2019: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:33:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:33:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:33:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:33:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:33:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:25 vm11.local ceph-mon[53973]: pgmap v2020: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:33:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:33:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:33:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:33:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:33:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:25 vm10.local ceph-mon[53712]: pgmap v2020: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:27 vm11.local ceph-mon[53973]: pgmap v2021: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:28.169 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:28.170 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:27 vm10.local ceph-mon[53712]: pgmap v2021: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:28.196 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:28.196 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:33:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:33:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:33:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:33:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:29 vm10.local ceph-mon[53712]: pgmap v2022: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:29 vm11.local ceph-mon[53973]: pgmap v2022: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:31 vm10.local ceph-mon[53712]: pgmap v2023: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:31 vm11.local ceph-mon[53973]: pgmap v2023: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:33.198 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:33.198 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:33.223 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:33.224 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:34.142 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:33 vm11.local ceph-mon[53973]: pgmap v2024: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:33 vm10.local ceph-mon[53712]: pgmap v2024: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:35 vm11.local ceph-mon[53973]: pgmap v2025: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:35 vm10.local ceph-mon[53712]: pgmap v2025: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:37 vm11.local ceph-mon[53973]: pgmap v2026: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:37 vm10.local ceph-mon[53712]: pgmap v2026: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:38.225 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:38.225 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:38.250 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:38.250 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:39 vm10.local ceph-mon[53712]: pgmap v2027: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:39 vm11.local ceph-mon[53973]: pgmap v2027: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:41 vm10.local ceph-mon[53712]: pgmap v2028: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:41 vm11.local ceph-mon[53973]: pgmap v2028: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:43.252 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:43.252 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:43.277 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:43.277 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:43 vm10.local ceph-mon[53712]: pgmap v2029: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:43 vm11.local ceph-mon[53973]: pgmap v2029: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:45 vm11.local ceph-mon[53973]: pgmap v2030: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:45 vm10.local ceph-mon[53712]: pgmap v2030: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:47 vm11.local ceph-mon[53973]: pgmap v2031: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:47 vm10.local ceph-mon[53712]: pgmap v2031: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:48.279 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:48.279 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:48.305 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:48.305 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:49 vm10.local ceph-mon[53712]: pgmap v2032: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:49 vm11.local ceph-mon[53973]: pgmap v2032: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:51 vm10.local ceph-mon[53712]: pgmap v2033: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:51 vm11.local ceph-mon[53973]: pgmap v2033: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:53.307 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:53.307 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:53.332 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:53.333 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:33:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:53 vm10.local ceph-mon[53712]: pgmap v2034: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:53 vm11.local ceph-mon[53973]: pgmap v2034: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:55 vm11.local ceph-mon[53973]: pgmap v2035: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:55 vm10.local ceph-mon[53712]: pgmap v2035: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:33:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:57 vm11.local ceph-mon[53973]: pgmap v2036: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:57 vm10.local ceph-mon[53712]: pgmap v2036: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:33:58.334 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:33:58.334 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:33:58.360 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:33:58.360 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:33:59 vm10.local ceph-mon[53712]: pgmap v2037: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:33:59 vm11.local ceph-mon[53973]: pgmap v2037: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:01 vm10.local ceph-mon[53712]: pgmap v2038: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:01 vm11.local ceph-mon[53973]: pgmap v2038: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:03.362 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:03.362 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:03.388 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:03.388 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:04.141 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:03 vm11.local ceph-mon[53973]: pgmap v2039: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:03 vm10.local ceph-mon[53712]: pgmap v2039: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:05.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:05 vm10.local ceph-mon[53712]: pgmap v2040: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:06.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:05 vm11.local ceph-mon[53973]: pgmap v2040: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:07 vm11.local ceph-mon[53973]: pgmap v2041: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:07 vm10.local ceph-mon[53712]: pgmap v2041: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:08.390 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:08.390 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:08.416 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:08.416 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:09 vm10.local ceph-mon[53712]: pgmap v2042: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:09 vm11.local ceph-mon[53973]: pgmap v2042: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:11 vm10.local ceph-mon[53712]: pgmap v2043: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:11 vm11.local ceph-mon[53973]: pgmap v2043: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:13.418 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:13.418 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:13.444 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:13.444 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:13 vm10.local ceph-mon[53712]: pgmap v2044: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:13 vm11.local ceph-mon[53973]: pgmap v2044: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:15 vm11.local ceph-mon[53973]: pgmap v2045: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:15 vm10.local ceph-mon[53712]: pgmap v2045: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:17 vm11.local ceph-mon[53973]: pgmap v2046: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:17 vm10.local ceph-mon[53712]: pgmap v2046: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:18.445 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:18.446 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:18.471 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:18.472 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:19 vm10.local ceph-mon[53712]: pgmap v2047: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:19 vm11.local ceph-mon[53973]: pgmap v2047: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:21 vm10.local ceph-mon[53712]: pgmap v2048: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:21 vm11.local ceph-mon[53973]: pgmap v2048: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:23.473 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:23.474 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:23.500 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:23.500 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:23 vm10.local ceph-mon[53712]: pgmap v2049: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:23 vm11.local ceph-mon[53973]: pgmap v2049: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:34:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:34:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:34:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:25 vm11.local ceph-mon[53973]: pgmap v2050: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:34:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:34:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:34:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:34:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:34:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:25 vm10.local ceph-mon[53712]: pgmap v2050: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:34:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:34:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:27 vm11.local ceph-mon[53973]: pgmap v2051: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:27 vm10.local ceph-mon[53712]: pgmap v2051: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:28.501 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:28.502 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:28.533 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:28.533 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:34:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:34:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:34:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:34:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:29 vm10.local ceph-mon[53712]: pgmap v2052: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:29 vm11.local ceph-mon[53973]: pgmap v2052: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:31 vm10.local ceph-mon[53712]: pgmap v2053: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:31 vm11.local ceph-mon[53973]: pgmap v2053: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:33.534 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:33.535 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:33.562 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:33.563 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:33 vm10.local ceph-mon[53712]: pgmap v2054: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:33 vm11.local ceph-mon[53973]: pgmap v2054: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:35 vm11.local ceph-mon[53973]: pgmap v2055: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:35 vm10.local ceph-mon[53712]: pgmap v2055: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:36 vm11.local ceph-mon[53973]: pgmap v2056: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:36 vm10.local ceph-mon[53712]: pgmap v2056: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:38.564 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:38.564 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:38.591 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:38.591 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:39 vm10.local ceph-mon[53712]: pgmap v2057: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:39 vm11.local ceph-mon[53973]: pgmap v2057: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:41 vm10.local ceph-mon[53712]: pgmap v2058: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:41 vm11.local ceph-mon[53973]: pgmap v2058: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:43.592 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:43.593 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:43.618 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:43.619 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:43 vm10.local ceph-mon[53712]: pgmap v2059: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:43 vm11.local ceph-mon[53973]: pgmap v2059: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:45 vm11.local ceph-mon[53973]: pgmap v2060: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:45 vm10.local ceph-mon[53712]: pgmap v2060: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:47 vm11.local ceph-mon[53973]: pgmap v2061: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:47 vm10.local ceph-mon[53712]: pgmap v2061: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:48.620 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:48.620 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:48.646 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:48.647 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:49 vm10.local ceph-mon[53712]: pgmap v2062: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:49 vm11.local ceph-mon[53973]: pgmap v2062: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:51 vm10.local ceph-mon[53712]: pgmap v2063: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:51 vm11.local ceph-mon[53973]: pgmap v2063: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:53.648 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:53.648 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:53.674 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:53.674 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:34:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:53 vm10.local ceph-mon[53712]: pgmap v2064: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:53 vm11.local ceph-mon[53973]: pgmap v2064: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:55 vm11.local ceph-mon[53973]: pgmap v2065: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:55 vm10.local ceph-mon[53712]: pgmap v2065: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:34:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:57 vm11.local ceph-mon[53973]: pgmap v2066: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:57 vm10.local ceph-mon[53712]: pgmap v2066: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:34:58.675 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:34:58.676 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:34:58.702 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:34:58.702 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:34:59 vm10.local ceph-mon[53712]: pgmap v2067: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:34:59 vm11.local ceph-mon[53973]: pgmap v2067: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:01 vm10.local ceph-mon[53712]: pgmap v2068: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:01 vm11.local ceph-mon[53973]: pgmap v2068: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:03.704 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:03.704 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:03.729 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:03.730 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:03 vm10.local ceph-mon[53712]: pgmap v2069: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:04.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:03 vm11.local ceph-mon[53973]: pgmap v2069: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:05 vm11.local ceph-mon[53973]: pgmap v2070: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:05 vm10.local ceph-mon[53712]: pgmap v2070: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:08.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:08 vm10.local ceph-mon[53712]: pgmap v2071: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:08.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:08 vm11.local ceph-mon[53973]: pgmap v2071: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:08.731 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:08.731 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:08.757 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:08.758 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:09 vm10.local ceph-mon[53712]: pgmap v2072: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:09 vm11.local ceph-mon[53973]: pgmap v2072: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:11 vm10.local ceph-mon[53712]: pgmap v2073: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:11 vm11.local ceph-mon[53973]: pgmap v2073: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:13.759 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:13.759 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:13.785 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:13.785 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:13 vm10.local ceph-mon[53712]: pgmap v2074: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:13 vm11.local ceph-mon[53973]: pgmap v2074: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:15 vm11.local ceph-mon[53973]: pgmap v2075: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:15 vm10.local ceph-mon[53712]: pgmap v2075: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:17 vm11.local ceph-mon[53973]: pgmap v2076: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:17 vm10.local ceph-mon[53712]: pgmap v2076: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:18.787 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:18.787 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:18.813 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:18.813 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:19 vm10.local ceph-mon[53712]: pgmap v2077: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:19 vm11.local ceph-mon[53973]: pgmap v2077: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:21 vm10.local ceph-mon[53712]: pgmap v2078: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:21 vm11.local ceph-mon[53973]: pgmap v2078: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:23.815 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:23.815 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:23.840 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:23.841 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:23 vm10.local ceph-mon[53712]: pgmap v2079: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:23 vm11.local ceph-mon[53973]: pgmap v2079: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:25 vm11.local ceph-mon[53973]: pgmap v2080: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:35:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:35:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:35:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:35:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:35:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:25 vm10.local ceph-mon[53712]: pgmap v2080: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:35:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:35:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:35:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:35:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:35:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:27 vm11.local ceph-mon[53973]: pgmap v2081: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:27 vm10.local ceph-mon[53712]: pgmap v2081: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:28.842 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:28.842 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:28.870 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:28.871 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:35:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:35:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:35:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:35:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:29 vm10.local ceph-mon[53712]: pgmap v2082: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:29 vm11.local ceph-mon[53973]: pgmap v2082: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:31 vm10.local ceph-mon[53712]: pgmap v2083: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:31 vm11.local ceph-mon[53973]: pgmap v2083: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:33.872 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:33.872 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:33.903 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:33.903 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:33 vm10.local ceph-mon[53712]: pgmap v2084: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:33 vm11.local ceph-mon[53973]: pgmap v2084: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:35 vm11.local ceph-mon[53973]: pgmap v2085: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:35 vm10.local ceph-mon[53712]: pgmap v2085: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:37 vm11.local ceph-mon[53973]: pgmap v2086: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:37 vm10.local ceph-mon[53712]: pgmap v2086: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:38.905 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:38.905 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:38.934 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:38.934 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:39 vm10.local ceph-mon[53712]: pgmap v2087: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:39 vm11.local ceph-mon[53973]: pgmap v2087: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:41 vm10.local ceph-mon[53712]: pgmap v2088: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:41 vm11.local ceph-mon[53973]: pgmap v2088: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:43.935 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:43.936 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:43.961 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:43.961 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:43 vm10.local ceph-mon[53712]: pgmap v2089: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:43 vm11.local ceph-mon[53973]: pgmap v2089: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:45 vm11.local ceph-mon[53973]: pgmap v2090: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:45 vm10.local ceph-mon[53712]: pgmap v2090: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:48.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:47 vm11.local ceph-mon[53973]: pgmap v2091: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:47 vm10.local ceph-mon[53712]: pgmap v2091: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:48.962 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:48.963 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:48.990 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:48.991 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:49 vm10.local ceph-mon[53712]: pgmap v2092: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:49 vm11.local ceph-mon[53973]: pgmap v2092: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:51 vm10.local ceph-mon[53712]: pgmap v2093: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:51 vm11.local ceph-mon[53973]: pgmap v2093: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:53.992 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:53.992 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:54.019 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:54.019 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:35:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:53 vm10.local ceph-mon[53712]: pgmap v2094: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:53 vm11.local ceph-mon[53973]: pgmap v2094: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:55 vm11.local ceph-mon[53973]: pgmap v2095: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:55 vm10.local ceph-mon[53712]: pgmap v2095: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:35:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:57 vm10.local ceph-mon[53712]: pgmap v2096: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:58.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:57 vm11.local ceph-mon[53973]: pgmap v2096: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:35:59.020 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:35:59.021 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:35:59.047 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:35:59.048 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:35:59 vm10.local ceph-mon[53712]: pgmap v2097: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:35:59 vm11.local ceph-mon[53973]: pgmap v2097: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:00 vm11.local ceph-mon[53973]: pgmap v2098: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:00 vm10.local ceph-mon[53712]: pgmap v2098: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:04.049 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:04.050 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:04.075 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:04.076 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:03 vm10.local ceph-mon[53712]: pgmap v2099: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:04.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:03 vm11.local ceph-mon[53973]: pgmap v2099: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:05 vm11.local ceph-mon[53973]: pgmap v2100: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:05 vm10.local ceph-mon[53712]: pgmap v2100: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:07 vm11.local ceph-mon[53973]: pgmap v2101: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:07 vm10.local ceph-mon[53712]: pgmap v2101: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:09.077 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:09.078 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:09.105 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:09.106 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:09 vm10.local ceph-mon[53712]: pgmap v2102: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:09 vm11.local ceph-mon[53973]: pgmap v2102: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:11 vm10.local ceph-mon[53712]: pgmap v2103: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:11 vm11.local ceph-mon[53973]: pgmap v2103: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:14.107 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:14.108 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:14.133 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:14.133 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:13 vm10.local ceph-mon[53712]: pgmap v2104: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:13 vm11.local ceph-mon[53973]: pgmap v2104: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:15 vm11.local ceph-mon[53973]: pgmap v2105: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:15 vm10.local ceph-mon[53712]: pgmap v2105: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:17 vm11.local ceph-mon[53973]: pgmap v2106: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:17 vm10.local ceph-mon[53712]: pgmap v2106: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:19.135 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:19.135 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:19.160 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:19.161 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:19 vm10.local ceph-mon[53712]: pgmap v2107: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:19 vm11.local ceph-mon[53973]: pgmap v2107: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:21 vm10.local ceph-mon[53712]: pgmap v2108: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:21 vm11.local ceph-mon[53973]: pgmap v2108: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:24.162 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:24.163 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:23 vm10.local ceph-mon[53712]: pgmap v2109: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:24.189 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:24.190 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:23 vm11.local ceph-mon[53973]: pgmap v2109: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:25.550 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:25 vm11.local ceph-mon[53973]: pgmap v2110: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:25.550 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:36:25.550 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:36:25.550 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:36:25.550 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:36:25.550 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:36:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:25 vm10.local ceph-mon[53712]: pgmap v2110: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:36:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:36:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:36:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:36:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:36:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:27 vm11.local ceph-mon[53973]: pgmap v2111: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:27 vm10.local ceph-mon[53712]: pgmap v2111: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:36:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:36:29.191 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:29.192 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:29.218 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:29.218 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:36:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:36:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:29 vm10.local ceph-mon[53712]: pgmap v2112: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:29 vm11.local ceph-mon[53973]: pgmap v2112: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:31 vm10.local ceph-mon[53712]: pgmap v2113: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:31 vm11.local ceph-mon[53973]: pgmap v2113: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:33 vm10.local ceph-mon[53712]: pgmap v2114: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:34.220 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:34.220 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:34.247 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:34.247 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:33 vm11.local ceph-mon[53973]: pgmap v2114: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:35 vm11.local ceph-mon[53973]: pgmap v2115: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:35 vm10.local ceph-mon[53712]: pgmap v2115: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:38.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:37 vm11.local ceph-mon[53973]: pgmap v2116: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:37 vm10.local ceph-mon[53712]: pgmap v2116: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:39.248 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:39.249 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:39.278 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:39.279 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:39 vm10.local ceph-mon[53712]: pgmap v2117: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:39 vm11.local ceph-mon[53973]: pgmap v2117: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:41 vm10.local ceph-mon[53712]: pgmap v2118: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:41 vm11.local ceph-mon[53973]: pgmap v2118: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:43 vm10.local ceph-mon[53712]: pgmap v2119: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:44.280 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:44.281 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:44.309 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:44.309 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:43 vm11.local ceph-mon[53973]: pgmap v2119: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:45 vm11.local ceph-mon[53973]: pgmap v2120: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:45 vm10.local ceph-mon[53712]: pgmap v2120: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:47 vm10.local ceph-mon[53712]: pgmap v2121: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:47 vm11.local ceph-mon[53973]: pgmap v2121: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:49.310 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:49.311 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:49.336 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:49.337 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:49 vm10.local ceph-mon[53712]: pgmap v2122: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:49 vm11.local ceph-mon[53973]: pgmap v2122: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:51 vm10.local ceph-mon[53712]: pgmap v2123: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:51 vm11.local ceph-mon[53973]: pgmap v2123: 97 pgs: 97 active+clean; 453 KiB data, 73 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:53 vm10.local ceph-mon[53712]: pgmap v2124: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:54.338 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:54.339 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:53 vm11.local ceph-mon[53973]: pgmap v2124: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:54.365 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:54.366 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:36:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:55 vm11.local ceph-mon[53973]: pgmap v2125: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:55 vm10.local ceph-mon[53712]: pgmap v2125: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:36:58.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:57 vm11.local ceph-mon[53973]: pgmap v2126: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:57 vm10.local ceph-mon[53712]: pgmap v2126: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:36:59.367 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:36:59.368 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:36:59.394 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:36:59.395 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:36:59 vm10.local ceph-mon[53712]: pgmap v2127: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:36:59 vm11.local ceph-mon[53973]: pgmap v2127: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:01 vm10.local ceph-mon[53712]: pgmap v2128: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:01 vm11.local ceph-mon[53973]: pgmap v2128: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:03 vm10.local ceph-mon[53712]: pgmap v2129: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:04.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:03 vm11.local ceph-mon[53973]: pgmap v2129: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:04.396 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:04.397 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:04.422 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:04.422 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:05 vm11.local ceph-mon[53973]: pgmap v2130: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:05 vm10.local ceph-mon[53712]: pgmap v2130: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:08.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:07 vm11.local ceph-mon[53973]: pgmap v2131: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:07 vm10.local ceph-mon[53712]: pgmap v2131: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:09.424 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:09.424 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:09.451 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:09.451 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:09 vm10.local ceph-mon[53712]: pgmap v2132: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:09 vm11.local ceph-mon[53973]: pgmap v2132: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:11 vm10.local ceph-mon[53712]: pgmap v2133: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:11 vm11.local ceph-mon[53973]: pgmap v2133: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:13 vm10.local ceph-mon[53712]: pgmap v2134: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:13 vm11.local ceph-mon[53973]: pgmap v2134: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:14.452 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:14.453 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:14.479 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:14.480 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:15 vm11.local ceph-mon[53973]: pgmap v2135: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:15 vm10.local ceph-mon[53712]: pgmap v2135: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:18.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:17 vm11.local ceph-mon[53973]: pgmap v2136: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:18.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:17 vm10.local ceph-mon[53712]: pgmap v2136: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:19.481 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:19.482 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:19.508 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:19.508 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:20.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:19 vm10.local ceph-mon[53712]: pgmap v2137: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:19 vm11.local ceph-mon[53973]: pgmap v2137: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:21 vm10.local ceph-mon[53712]: pgmap v2138: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:21 vm11.local ceph-mon[53973]: pgmap v2138: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:23 vm10.local ceph-mon[53712]: pgmap v2139: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:23 vm11.local ceph-mon[53973]: pgmap v2139: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:24.509 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:24.510 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:24.537 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:24.537 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:25 vm11.local ceph-mon[53973]: pgmap v2140: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:37:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:37:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:25 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:37:25.882 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:25 vm10.local ceph-mon[53712]: pgmap v2140: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:25.882 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:37:25.882 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:37:25.882 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:25 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:37:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:37:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:37:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:37:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:37:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:37:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:37:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:27 vm10.local ceph-mon[53712]: pgmap v2141: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:27 vm11.local ceph-mon[53973]: pgmap v2141: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:37:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:37:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:37:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:37:29.539 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:29.539 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:29.567 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:29.568 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:29 vm10.local ceph-mon[53712]: pgmap v2142: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:29 vm11.local ceph-mon[53973]: pgmap v2142: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:31 vm10.local ceph-mon[53712]: pgmap v2143: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:31 vm11.local ceph-mon[53973]: pgmap v2143: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:33 vm10.local ceph-mon[53712]: pgmap v2144: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:33 vm11.local ceph-mon[53973]: pgmap v2144: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:34.569 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:34.570 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:34.597 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:34.598 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:35.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:35 vm11.local ceph-mon[53973]: pgmap v2145: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:35 vm10.local ceph-mon[53712]: pgmap v2145: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:37 vm10.local ceph-mon[53712]: pgmap v2146: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:38.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:37 vm11.local ceph-mon[53973]: pgmap v2146: 97 pgs: 97 active+clean; 453 KiB data, 77 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:39.599 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:39.599 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:39.625 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:39.626 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:39 vm10.local ceph-mon[53712]: pgmap v2147: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:39 vm11.local ceph-mon[53973]: pgmap v2147: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:42.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:41 vm10.local ceph-mon[53712]: pgmap v2148: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:41 vm11.local ceph-mon[53973]: pgmap v2148: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:43 vm10.local ceph-mon[53712]: pgmap v2149: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:43 vm11.local ceph-mon[53973]: pgmap v2149: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:44.627 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:44.628 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:44.653 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:44.653 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:45.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:45 vm11.local ceph-mon[53973]: pgmap v2150: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:45 vm10.local ceph-mon[53712]: pgmap v2150: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:47 vm10.local ceph-mon[53712]: pgmap v2151: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:48.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:47 vm11.local ceph-mon[53973]: pgmap v2151: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:49.654 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:49.655 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:49.681 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:49.681 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:50.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:49 vm10.local ceph-mon[53712]: pgmap v2152: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:49 vm11.local ceph-mon[53973]: pgmap v2152: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:52.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:51 vm10.local ceph-mon[53712]: pgmap v2153: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:51 vm11.local ceph-mon[53973]: pgmap v2153: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:53 vm10.local ceph-mon[53712]: pgmap v2154: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:53 vm11.local ceph-mon[53973]: pgmap v2154: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:54.683 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:54.684 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:54.709 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:54.710 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:37:55.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:55 vm11.local ceph-mon[53973]: pgmap v2155: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:55 vm10.local ceph-mon[53712]: pgmap v2155: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:37:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:57 vm10.local ceph-mon[53712]: pgmap v2156: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:58.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:57 vm11.local ceph-mon[53973]: pgmap v2156: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:37:59.711 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:37:59.712 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:37:59.739 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:37:59.740 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:37:59 vm10.local ceph-mon[53712]: pgmap v2157: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:37:59 vm11.local ceph-mon[53973]: pgmap v2157: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:02.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:01 vm10.local ceph-mon[53712]: pgmap v2158: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:01 vm11.local ceph-mon[53973]: pgmap v2158: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:03 vm10.local ceph-mon[53712]: pgmap v2159: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:04.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:03 vm11.local ceph-mon[53973]: pgmap v2159: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:04.741 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:04.742 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:04.768 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:04.768 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:05.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:05 vm11.local ceph-mon[53973]: pgmap v2160: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:05 vm10.local ceph-mon[53712]: pgmap v2160: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:07 vm10.local ceph-mon[53712]: pgmap v2161: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:08.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:07 vm11.local ceph-mon[53973]: pgmap v2161: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:09.770 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:09.770 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:09.796 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:09.796 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:10.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:09 vm10.local ceph-mon[53712]: pgmap v2162: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:09 vm11.local ceph-mon[53973]: pgmap v2162: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:12.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:11 vm10.local ceph-mon[53712]: pgmap v2163: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:11 vm11.local ceph-mon[53973]: pgmap v2163: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:13 vm10.local ceph-mon[53712]: pgmap v2164: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:13 vm11.local ceph-mon[53973]: pgmap v2164: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:14.797 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:14.797 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:14.823 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:14.824 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:15.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:15 vm11.local ceph-mon[53973]: pgmap v2165: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:15 vm10.local ceph-mon[53712]: pgmap v2165: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:18.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:18 vm10.local ceph-mon[53712]: pgmap v2166: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:18 vm11.local ceph-mon[53973]: pgmap v2166: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:19.337 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:19 vm10.local ceph-mon[53712]: pgmap v2167: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:19 vm11.local ceph-mon[53973]: pgmap v2167: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:19.825 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:19.826 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:19.851 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:19.852 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:22.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:21 vm10.local ceph-mon[53712]: pgmap v2168: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:21 vm11.local ceph-mon[53973]: pgmap v2168: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:23 vm10.local ceph-mon[53712]: pgmap v2169: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:23 vm11.local ceph-mon[53973]: pgmap v2169: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:24.855 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:24.855 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:24.883 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:24.883 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:25.955 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:25 vm10.local ceph-mon[53712]: pgmap v2170: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:26.084 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:25 vm11.local ceph-mon[53973]: pgmap v2170: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:38:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:38:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:38:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:38:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:38:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:38:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:38:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:38:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:38:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:38:27.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:38:27.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:38:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:27 vm11.local ceph-mon[53973]: pgmap v2171: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:28.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:27 vm10.local ceph-mon[53712]: pgmap v2171: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:28.976 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:38:28.976 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:38:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:38:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:38:29.885 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:29.885 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:29.914 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:29.915 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:29 vm11.local ceph-mon[53973]: pgmap v2172: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:29 vm10.local ceph-mon[53712]: pgmap v2172: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:32.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:31 vm10.local ceph-mon[53712]: pgmap v2173: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:31 vm11.local ceph-mon[53973]: pgmap v2173: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:33 vm10.local ceph-mon[53712]: pgmap v2174: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:33 vm11.local ceph-mon[53973]: pgmap v2174: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:34.916 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:34.916 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:34.944 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:34.945 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:35 vm10.local ceph-mon[53712]: pgmap v2175: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:36.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:35 vm11.local ceph-mon[53973]: pgmap v2175: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:38.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:37 vm10.local ceph-mon[53712]: pgmap v2176: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:38.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:37 vm11.local ceph-mon[53973]: pgmap v2176: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:39.946 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:39.946 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:39.974 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:39.975 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:39 vm10.local ceph-mon[53712]: pgmap v2177: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:39 vm11.local ceph-mon[53973]: pgmap v2177: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:41 vm11.local ceph-mon[53973]: pgmap v2178: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:41 vm10.local ceph-mon[53712]: pgmap v2178: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:43 vm10.local ceph-mon[53712]: pgmap v2179: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:43 vm11.local ceph-mon[53973]: pgmap v2179: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:44.977 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:44.977 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:45.002 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:45.003 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:45 vm10.local ceph-mon[53712]: pgmap v2180: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:45 vm11.local ceph-mon[53973]: pgmap v2180: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:47 vm10.local ceph-mon[53712]: pgmap v2181: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:48.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:47 vm11.local ceph-mon[53973]: pgmap v2181: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:50.004 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:50.005 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:50.031 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:50.031 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:49 vm11.local ceph-mon[53973]: pgmap v2182: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:50.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:49 vm10.local ceph-mon[53712]: pgmap v2182: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:51 vm11.local ceph-mon[53973]: pgmap v2183: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:52.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:51 vm10.local ceph-mon[53712]: pgmap v2183: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:53 vm10.local ceph-mon[53712]: pgmap v2184: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:53 vm11.local ceph-mon[53973]: pgmap v2184: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:55.032 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:38:55.033 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:38:55.060 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:38:55.060 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:38:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:55 vm10.local ceph-mon[53712]: pgmap v2185: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:56.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:55 vm11.local ceph-mon[53973]: pgmap v2185: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:38:58.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:57 vm10.local ceph-mon[53712]: pgmap v2186: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:38:58.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:57 vm11.local ceph-mon[53973]: pgmap v2186: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:00.062 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:00.062 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:00.097 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:00.098 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:00.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:38:59 vm10.local ceph-mon[53712]: pgmap v2187: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:38:59 vm11.local ceph-mon[53973]: pgmap v2187: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:01 vm11.local ceph-mon[53973]: pgmap v2188: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:02.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:01 vm10.local ceph-mon[53712]: pgmap v2188: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:03 vm10.local ceph-mon[53712]: pgmap v2189: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:04.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:03 vm11.local ceph-mon[53973]: pgmap v2189: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:05.099 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:05.099 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:05.125 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:05.126 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:05 vm10.local ceph-mon[53712]: pgmap v2190: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:06.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:05 vm11.local ceph-mon[53973]: pgmap v2190: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:08.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:07 vm10.local ceph-mon[53712]: pgmap v2191: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:08.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:07 vm11.local ceph-mon[53973]: pgmap v2191: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:10.127 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:10.127 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:10.152 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:10.153 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:09 vm11.local ceph-mon[53973]: pgmap v2192: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:10.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:09 vm10.local ceph-mon[53712]: pgmap v2192: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:11 vm11.local ceph-mon[53973]: pgmap v2193: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:12.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:11 vm10.local ceph-mon[53712]: pgmap v2193: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:13 vm10.local ceph-mon[53712]: pgmap v2194: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:13 vm11.local ceph-mon[53973]: pgmap v2194: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:15.154 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:15.155 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:15.180 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:15.180 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:15 vm10.local ceph-mon[53712]: pgmap v2195: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:16.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:15 vm11.local ceph-mon[53973]: pgmap v2195: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:18.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:17 vm11.local ceph-mon[53973]: pgmap v2196: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:18.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:17 vm10.local ceph-mon[53712]: pgmap v2196: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:20.182 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:20.182 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:20.208 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:20.208 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:19 vm11.local ceph-mon[53973]: pgmap v2197: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:20.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:19 vm10.local ceph-mon[53712]: pgmap v2197: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:21 vm11.local ceph-mon[53973]: pgmap v2198: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:22.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:21 vm10.local ceph-mon[53712]: pgmap v2198: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:23 vm10.local ceph-mon[53712]: pgmap v2199: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:23 vm11.local ceph-mon[53973]: pgmap v2199: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:25.209 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:25.210 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:25.235 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:25.235 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:25 vm10.local ceph-mon[53712]: pgmap v2200: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:26.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:25 vm11.local ceph-mon[53973]: pgmap v2200: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:26.607 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:39:26.607 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:39:26.608 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:39:26.608 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:39:26.608 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:39:26.608 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:39:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:39:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:39:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:39:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:39:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:39:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:39:27.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:27 vm10.local ceph-mon[53712]: pgmap v2201: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:27 vm11.local ceph-mon[53973]: pgmap v2201: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:39:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:39:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:39:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:39:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:29 vm10.local ceph-mon[53712]: pgmap v2202: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:29 vm11.local ceph-mon[53973]: pgmap v2202: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:30.236 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:30.237 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:30.262 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:30.263 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:31 vm11.local ceph-mon[53973]: pgmap v2203: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:31 vm10.local ceph-mon[53712]: pgmap v2203: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:33 vm10.local ceph-mon[53712]: pgmap v2204: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:33 vm11.local ceph-mon[53973]: pgmap v2204: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:35.264 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:35.264 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:35.290 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:35.290 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:35 vm10.local ceph-mon[53712]: pgmap v2205: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:36.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:35 vm11.local ceph-mon[53973]: pgmap v2205: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:38.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:37 vm11.local ceph-mon[53973]: pgmap v2206: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:38.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:37 vm10.local ceph-mon[53712]: pgmap v2206: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:40.291 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:40.292 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:40.317 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:40.317 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:39 vm11.local ceph-mon[53973]: pgmap v2207: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:40.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:39 vm10.local ceph-mon[53712]: pgmap v2207: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:41 vm11.local ceph-mon[53973]: pgmap v2208: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:41 vm10.local ceph-mon[53712]: pgmap v2208: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:43 vm10.local ceph-mon[53712]: pgmap v2209: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:43 vm11.local ceph-mon[53973]: pgmap v2209: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:45.319 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:45.319 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:45.345 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:45.345 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:45 vm10.local ceph-mon[53712]: pgmap v2210: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:45 vm11.local ceph-mon[53973]: pgmap v2210: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:48.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:47 vm11.local ceph-mon[53973]: pgmap v2211: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:48.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:47 vm10.local ceph-mon[53712]: pgmap v2211: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:49 vm11.local ceph-mon[53973]: pgmap v2212: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:50.347 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:50.347 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:50.373 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:50.373 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:50.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:49 vm10.local ceph-mon[53712]: pgmap v2212: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:51 vm11.local ceph-mon[53973]: pgmap v2213: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:52.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:51 vm10.local ceph-mon[53712]: pgmap v2213: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:53 vm10.local ceph-mon[53712]: pgmap v2214: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:53 vm11.local ceph-mon[53973]: pgmap v2214: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:55.374 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:39:55.375 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:39:55.399 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:39:55.400 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:39:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:55 vm10.local ceph-mon[53712]: pgmap v2215: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:56.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:55 vm11.local ceph-mon[53973]: pgmap v2215: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:39:58.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:57 vm11.local ceph-mon[53973]: pgmap v2216: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:39:58.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:57 vm10.local ceph-mon[53712]: pgmap v2216: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:39:59 vm11.local ceph-mon[53973]: pgmap v2217: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:00.401 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:00.402 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:00.427 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:00.428 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:00.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:39:59 vm10.local ceph-mon[53712]: pgmap v2217: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:01 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T18:40:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:01 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T18:40:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:02 vm11.local ceph-mon[53973]: pgmap v2218: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:02.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:02 vm10.local ceph-mon[53712]: pgmap v2218: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:03 vm11.local ceph-mon[53973]: pgmap v2219: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:03.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:03 vm10.local ceph-mon[53712]: pgmap v2219: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:05.429 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:05.429 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:05.457 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:05.457 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:05 vm10.local ceph-mon[53712]: pgmap v2220: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:06.043 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:05 vm11.local ceph-mon[53973]: pgmap v2220: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:08.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:07 vm11.local ceph-mon[53973]: pgmap v2221: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:08.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:07 vm10.local ceph-mon[53712]: pgmap v2221: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:09 vm11.local ceph-mon[53973]: pgmap v2222: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:10.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:09 vm10.local ceph-mon[53712]: pgmap v2222: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:10.459 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:10.459 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:10.485 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:10.486 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:11 vm11.local ceph-mon[53973]: pgmap v2223: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:12.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:11 vm10.local ceph-mon[53712]: pgmap v2223: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:13 vm10.local ceph-mon[53712]: pgmap v2224: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:13 vm11.local ceph-mon[53973]: pgmap v2224: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:15.487 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:15.488 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:15.512 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:15.513 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:15 vm10.local ceph-mon[53712]: pgmap v2225: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:16.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:15 vm11.local ceph-mon[53973]: pgmap v2225: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:18.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:17 vm11.local ceph-mon[53973]: pgmap v2226: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:18.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:17 vm10.local ceph-mon[53712]: pgmap v2226: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:19 vm11.local ceph-mon[53973]: pgmap v2227: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:20.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:19 vm10.local ceph-mon[53712]: pgmap v2227: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:20.514 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:20.514 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:20.542 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:20.542 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:21 vm11.local ceph-mon[53973]: pgmap v2228: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:22.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:21 vm10.local ceph-mon[53712]: pgmap v2228: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:23 vm10.local ceph-mon[53712]: pgmap v2229: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:23 vm11.local ceph-mon[53973]: pgmap v2229: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:25.543 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:25.543 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:25.570 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:25.571 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:25 vm10.local ceph-mon[53712]: pgmap v2230: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:26.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:25 vm11.local ceph-mon[53973]: pgmap v2230: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:40:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:40:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:26 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:40:26.858 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:40:26.859 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:40:26.859 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:26 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:40:27.732 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:27 vm10.local ceph-mon[53712]: pgmap v2231: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:27.732 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:27 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:40:27.732 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:27 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:40:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:27 vm11.local ceph-mon[53973]: pgmap v2231: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:27 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:40:28.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:27 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:40:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:40:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:40:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:40:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:40:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:40:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:40:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:40:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:40:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:40:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:40:30.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:29 vm11.local ceph-mon[53973]: pgmap v2232: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:30.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:29 vm10.local ceph-mon[53712]: pgmap v2232: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:30.572 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:30.573 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:30.599 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:30.600 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:31 vm11.local ceph-mon[53973]: pgmap v2233: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:31 vm10.local ceph-mon[53712]: pgmap v2233: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:33 vm10.local ceph-mon[53712]: pgmap v2234: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:33 vm11.local ceph-mon[53973]: pgmap v2234: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:35.601 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:35.602 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:35.634 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:35.634 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:35 vm10.local ceph-mon[53712]: pgmap v2235: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:36.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:35 vm11.local ceph-mon[53973]: pgmap v2235: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:38.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:37 vm11.local ceph-mon[53973]: pgmap v2236: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:38.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:37 vm10.local ceph-mon[53712]: pgmap v2236: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:39 vm11.local ceph-mon[53973]: pgmap v2237: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:40.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:39 vm10.local ceph-mon[53712]: pgmap v2237: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:40.635 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:40.636 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:40.662 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:40.662 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:42.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:41 vm11.local ceph-mon[53973]: pgmap v2238: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:41 vm10.local ceph-mon[53712]: pgmap v2238: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:43 vm11.local ceph-mon[53973]: pgmap v2239: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:43 vm10.local ceph-mon[53712]: pgmap v2239: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:45.663 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:45.664 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:45.689 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:45.689 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:45 vm10.local ceph-mon[53712]: pgmap v2240: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:45 vm11.local ceph-mon[53973]: pgmap v2240: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:48.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:47 vm11.local ceph-mon[53973]: pgmap v2241: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:48.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:47 vm10.local ceph-mon[53712]: pgmap v2241: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:50.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:49 vm11.local ceph-mon[53973]: pgmap v2242: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:50.345 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:49 vm10.local ceph-mon[53712]: pgmap v2242: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:50.691 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:50.691 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:50.729 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:50.730 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:51 vm11.local ceph-mon[53973]: pgmap v2243: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:52.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:51 vm10.local ceph-mon[53712]: pgmap v2243: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:53 vm10.local ceph-mon[53712]: pgmap v2244: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:53 vm11.local ceph-mon[53973]: pgmap v2244: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:55.731 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:40:55.732 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:40:55.759 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:40:55.760 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:40:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:55 vm10.local ceph-mon[53712]: pgmap v2245: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:56.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:55 vm11.local ceph-mon[53973]: pgmap v2245: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:40:58.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:57 vm11.local ceph-mon[53973]: pgmap v2246: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:40:58.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:57 vm10.local ceph-mon[53712]: pgmap v2246: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:40:59 vm11.local ceph-mon[53973]: pgmap v2247: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:00.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:40:59 vm10.local ceph-mon[53712]: pgmap v2247: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:00.762 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:00.762 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:00.790 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:00.791 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:01 vm11.local ceph-mon[53973]: pgmap v2248: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:02.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:01 vm10.local ceph-mon[53712]: pgmap v2248: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:03 vm11.local ceph-mon[53973]: pgmap v2249: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:03 vm10.local ceph-mon[53712]: pgmap v2249: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:05.792 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:05.793 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:05.819 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:05.820 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:05.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:05 vm10.local ceph-mon[53712]: pgmap v2250: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:06.043 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:05 vm11.local ceph-mon[53973]: pgmap v2250: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:08.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:07 vm11.local ceph-mon[53973]: pgmap v2251: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:08.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:07 vm10.local ceph-mon[53712]: pgmap v2251: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:09 vm11.local ceph-mon[53973]: pgmap v2252: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:10.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:09 vm10.local ceph-mon[53712]: pgmap v2252: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:10.822 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:10.822 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:10.851 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:10.851 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:12.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:11 vm11.local ceph-mon[53973]: pgmap v2253: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:12.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:11 vm10.local ceph-mon[53712]: pgmap v2253: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:12 vm11.local ceph-mon[53973]: pgmap v2254: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:12 vm10.local ceph-mon[53712]: pgmap v2254: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:15.852 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:15.853 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:15.881 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:15.882 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:15.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:15 vm10.local ceph-mon[53712]: pgmap v2255: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:16.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:15 vm11.local ceph-mon[53973]: pgmap v2255: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:18.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:17 vm11.local ceph-mon[53973]: pgmap v2256: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:18.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:17 vm10.local ceph-mon[53712]: pgmap v2256: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:20.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:19 vm11.local ceph-mon[53973]: pgmap v2257: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:20.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:19 vm10.local ceph-mon[53712]: pgmap v2257: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:20.884 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:20.884 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:20.912 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:20.913 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:21 vm11.local ceph-mon[53973]: pgmap v2258: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:22.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:21 vm10.local ceph-mon[53712]: pgmap v2258: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:22 vm11.local ceph-mon[53973]: pgmap v2259: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:22 vm10.local ceph-mon[53712]: pgmap v2259: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:25.915 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:25.915 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:25 vm10.local ceph-mon[53712]: pgmap v2260: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:25.941 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:25.942 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:26.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:25 vm11.local ceph-mon[53973]: pgmap v2260: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:28.244 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:27 vm11.local ceph-mon[53973]: pgmap v2261: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:28.273 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:27 vm10.local ceph-mon[53712]: pgmap v2261: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:41:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:41:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:41:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:41:29.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:41:29.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:41:29.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:41:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:41:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:41:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:41:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:41:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:41:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:41:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:41:30.097 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:29 vm10.local ceph-mon[53712]: pgmap v2262: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:30.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:29 vm11.local ceph-mon[53973]: pgmap v2262: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:30.944 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:30.944 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:30.971 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:30.971 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:31 vm11.local ceph-mon[53973]: pgmap v2263: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:31 vm10.local ceph-mon[53712]: pgmap v2263: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:32 vm11.local ceph-mon[53973]: pgmap v2264: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:32 vm10.local ceph-mon[53712]: pgmap v2264: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:35.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:35 vm10.local ceph-mon[53712]: pgmap v2265: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:35.973 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:35.974 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:36.000 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:36.001 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:36.043 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:35 vm11.local ceph-mon[53973]: pgmap v2265: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:38.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:37 vm11.local ceph-mon[53973]: pgmap v2266: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:38.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:37 vm10.local ceph-mon[53712]: pgmap v2266: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:39 vm11.local ceph-mon[53973]: pgmap v2267: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:40.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:39 vm10.local ceph-mon[53712]: pgmap v2267: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:41.002 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:41.003 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:41.032 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:41.032 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:40 vm11.local ceph-mon[53973]: pgmap v2268: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:40 vm10.local ceph-mon[53712]: pgmap v2268: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:43 vm10.local ceph-mon[53712]: pgmap v2269: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:43 vm11.local ceph-mon[53973]: pgmap v2269: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:45.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:45 vm10.local ceph-mon[53712]: pgmap v2270: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:46.034 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:46.034 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:46.063 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:46.064 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:45 vm11.local ceph-mon[53973]: pgmap v2270: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:48.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:47 vm11.local ceph-mon[53973]: pgmap v2271: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:48.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:47 vm10.local ceph-mon[53712]: pgmap v2271: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:49.336 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:49 vm10.local ceph-mon[53712]: pgmap v2272: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:49 vm11.local ceph-mon[53973]: pgmap v2272: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:51.066 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:51.066 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:51.092 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:51.093 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:51 vm11.local ceph-mon[53973]: pgmap v2273: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:52.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:51 vm10.local ceph-mon[53712]: pgmap v2273: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:54.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:53 vm10.local ceph-mon[53712]: pgmap v2274: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:54.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:53 vm11.local ceph-mon[53973]: pgmap v2274: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:55.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:55 vm10.local ceph-mon[53712]: pgmap v2275: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:56.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:55 vm11.local ceph-mon[53973]: pgmap v2275: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:41:56.094 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:41:56.095 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:41:56.122 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:41:56.122 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:41:58.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:57 vm11.local ceph-mon[53973]: pgmap v2276: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:41:58.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:57 vm10.local ceph-mon[53712]: pgmap v2276: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:41:59 vm11.local ceph-mon[53973]: pgmap v2277: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:00.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:41:59 vm10.local ceph-mon[53712]: pgmap v2277: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:01.123 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:01.124 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:01.151 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:01.152 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:00 vm11.local ceph-mon[53973]: pgmap v2278: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:00 vm10.local ceph-mon[53712]: pgmap v2278: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:04.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:03 vm10.local ceph-mon[53712]: pgmap v2279: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:04.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:03 vm11.local ceph-mon[53973]: pgmap v2279: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:06.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:05 vm11.local ceph-mon[53973]: pgmap v2280: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:06.153 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:06.154 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:06.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:05 vm10.local ceph-mon[53712]: pgmap v2280: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:06.185 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:06.185 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:07 vm11.local ceph-mon[53973]: pgmap v2281: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:07 vm10.local ceph-mon[53712]: pgmap v2281: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:10.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:09 vm11.local ceph-mon[53973]: pgmap v2282: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:10.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:09 vm10.local ceph-mon[53712]: pgmap v2282: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:11.187 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:11.187 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:11.213 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:11.214 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:11 vm11.local ceph-mon[53973]: pgmap v2283: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:11 vm10.local ceph-mon[53712]: pgmap v2283: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:14.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:13 vm10.local ceph-mon[53712]: pgmap v2284: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:14.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:13 vm11.local ceph-mon[53973]: pgmap v2284: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:16.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:15 vm11.local ceph-mon[53973]: pgmap v2285: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:16.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:15 vm10.local ceph-mon[53712]: pgmap v2285: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:16.216 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:16.216 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:16.243 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:16.243 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:18.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:17 vm11.local ceph-mon[53973]: pgmap v2286: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:18.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:17 vm10.local ceph-mon[53712]: pgmap v2286: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:19.336 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:19 vm10.local ceph-mon[53712]: pgmap v2287: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:19 vm11.local ceph-mon[53973]: pgmap v2287: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:21.245 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:21.245 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:21.272 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:21.273 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:22.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:21 vm11.local ceph-mon[53973]: pgmap v2288: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:22.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:21 vm10.local ceph-mon[53712]: pgmap v2288: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:24.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:23 vm10.local ceph-mon[53712]: pgmap v2289: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:24.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:23 vm11.local ceph-mon[53973]: pgmap v2289: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:26.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:25 vm11.local ceph-mon[53973]: pgmap v2290: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:26.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:25 vm10.local ceph-mon[53712]: pgmap v2290: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:26.274 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:26.275 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:26.301 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:26.301 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:27 vm11.local ceph-mon[53973]: pgmap v2291: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:27 vm10.local ceph-mon[53712]: pgmap v2291: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:28.575 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:42:28.575 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:42:28.575 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:42:28.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:42:28.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:42:28.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:42:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:42:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:42:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:42:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:42:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:29 vm10.local ceph-mon[53712]: pgmap v2292: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:42:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:42:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:42:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:42:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:29 vm11.local ceph-mon[53973]: pgmap v2292: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:31.302 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:31.303 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:31.329 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:31.329 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:31 vm11.local ceph-mon[53973]: pgmap v2293: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:31 vm10.local ceph-mon[53712]: pgmap v2293: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:33 vm11.local ceph-mon[53973]: pgmap v2294: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:33 vm10.local ceph-mon[53712]: pgmap v2294: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:36.044 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:35 vm11.local ceph-mon[53973]: pgmap v2295: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:36.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:35 vm10.local ceph-mon[53712]: pgmap v2295: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:36.331 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:36.332 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:36.358 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:36.359 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:37 vm11.local ceph-mon[53973]: pgmap v2296: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:37 vm10.local ceph-mon[53712]: pgmap v2296: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:40.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:39 vm11.local ceph-mon[53973]: pgmap v2297: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:40.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:39 vm10.local ceph-mon[53712]: pgmap v2297: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:41 vm11.local ceph-mon[53973]: pgmap v2298: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:41.363 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:41.363 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:41 vm10.local ceph-mon[53712]: pgmap v2298: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:41.696 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:41.697 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:43 vm10.local ceph-mon[53712]: pgmap v2299: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:43 vm11.local ceph-mon[53973]: pgmap v2299: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:45 vm11.local ceph-mon[53973]: pgmap v2300: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:42:46.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:45 vm10.local ceph-mon[53712]: pgmap v2300: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:42:46.698 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:46.699 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:46.725 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:46.726 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:48.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:47 vm11.local ceph-mon[53973]: pgmap v2301: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:48.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:47 vm10.local ceph-mon[53712]: pgmap v2301: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:49.336 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:49 vm10.local ceph-mon[53712]: pgmap v2302: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:42:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:49 vm11.local ceph-mon[53973]: pgmap v2302: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:42:51.727 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:51.728 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:51.754 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:51.754 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:52.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:51 vm11.local ceph-mon[53973]: pgmap v2303: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:42:52.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:51 vm10.local ceph-mon[53712]: pgmap v2303: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:42:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:52 vm11.local ceph-mon[53973]: pgmap v2304: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:52 vm10.local ceph-mon[53712]: pgmap v2304: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:56.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:55 vm11.local ceph-mon[53973]: pgmap v2305: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:56.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:55 vm10.local ceph-mon[53712]: pgmap v2305: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:56.756 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:42:56.756 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:42:56.956 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:42:56.956 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:42:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:57 vm11.local ceph-mon[53973]: pgmap v2306: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:57 vm10.local ceph-mon[53712]: pgmap v2306: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:42:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:42:59 vm11.local ceph-mon[53973]: pgmap v2307: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:42:59.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:42:59 vm10.local ceph-mon[53712]: pgmap v2307: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:01.958 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:01.959 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:01.986 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:01.986 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:02.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:01 vm11.local ceph-mon[53973]: pgmap v2308: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:02.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:01 vm10.local ceph-mon[53712]: pgmap v2308: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:02 vm11.local ceph-mon[53973]: pgmap v2309: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:02 vm10.local ceph-mon[53712]: pgmap v2309: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:06.044 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:05 vm11.local ceph-mon[53973]: pgmap v2310: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:06.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:05 vm10.local ceph-mon[53712]: pgmap v2310: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:06.988 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:06.988 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:07.016 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:07.017 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:07 vm11.local ceph-mon[53973]: pgmap v2311: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:07.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:07 vm10.local ceph-mon[53712]: pgmap v2311: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:09 vm11.local ceph-mon[53973]: pgmap v2312: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:09 vm10.local ceph-mon[53712]: pgmap v2312: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:11 vm11.local ceph-mon[53973]: pgmap v2313: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:11 vm10.local ceph-mon[53712]: pgmap v2313: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:12.018 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:12.019 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:12.047 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:12.048 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:13 vm11.local ceph-mon[53973]: pgmap v2314: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:13 vm10.local ceph-mon[53712]: pgmap v2314: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:16.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:15 vm11.local ceph-mon[53973]: pgmap v2315: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:16.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:15 vm10.local ceph-mon[53712]: pgmap v2315: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:17.049 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:17.050 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:17.076 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:17.077 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:17 vm11.local ceph-mon[53973]: pgmap v2316: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:17 vm10.local ceph-mon[53712]: pgmap v2316: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:19.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:19 vm10.local ceph-mon[53712]: pgmap v2317: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:19 vm11.local ceph-mon[53973]: pgmap v2317: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:21 vm11.local ceph-mon[53973]: pgmap v2318: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:21 vm10.local ceph-mon[53712]: pgmap v2318: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:22.078 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:22.079 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:22.105 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:22.106 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:23 vm11.local ceph-mon[53973]: pgmap v2319: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:23 vm10.local ceph-mon[53712]: pgmap v2319: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:26.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:25 vm11.local ceph-mon[53973]: pgmap v2320: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:26.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:25 vm10.local ceph-mon[53712]: pgmap v2320: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:27.107 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:27.108 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:27.138 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:27.139 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:27 vm11.local ceph-mon[53973]: pgmap v2321: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:27 vm10.local ceph-mon[53712]: pgmap v2321: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:28.828 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:43:28.828 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:43:28.846 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:43:28.846 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:43:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:43:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:43:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:43:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:43:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:43:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:29 vm11.local ceph-mon[53973]: pgmap v2322: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:43:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:43:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:43:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:43:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:43:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:29 vm10.local ceph-mon[53712]: pgmap v2322: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:32.140 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:32.140 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:32.166 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:32.166 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:32.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:31 vm11.local ceph-mon[53973]: pgmap v2323: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:31 vm10.local ceph-mon[53712]: pgmap v2323: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:32 vm11.local ceph-mon[53973]: pgmap v2324: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:32 vm10.local ceph-mon[53712]: pgmap v2324: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:36.043 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:35 vm11.local ceph-mon[53973]: pgmap v2325: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:36.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:35 vm10.local ceph-mon[53712]: pgmap v2325: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:37.168 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:37.168 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:37.194 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:37.194 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:37 vm11.local ceph-mon[53973]: pgmap v2326: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:37 vm10.local ceph-mon[53712]: pgmap v2326: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:39 vm11.local ceph-mon[53973]: pgmap v2327: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:39 vm10.local ceph-mon[53712]: pgmap v2327: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:41 vm11.local ceph-mon[53973]: pgmap v2328: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:41 vm10.local ceph-mon[53712]: pgmap v2328: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:42.196 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:42.197 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:42.225 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:42.226 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:43 vm11.local ceph-mon[53973]: pgmap v2329: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:43 vm10.local ceph-mon[53712]: pgmap v2329: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:45 vm11.local ceph-mon[53973]: pgmap v2330: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:46.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:45 vm10.local ceph-mon[53712]: pgmap v2330: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:47.227 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:47.228 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:47.256 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:47.257 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:47 vm11.local ceph-mon[53973]: pgmap v2331: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:47 vm10.local ceph-mon[53712]: pgmap v2331: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:49 vm10.local ceph-mon[53712]: pgmap v2332: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:49 vm11.local ceph-mon[53973]: pgmap v2332: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:51 vm11.local ceph-mon[53973]: pgmap v2333: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:51 vm10.local ceph-mon[53712]: pgmap v2333: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:52.259 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:52.259 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:52.287 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:52.288 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:53 vm11.local ceph-mon[53973]: pgmap v2334: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:53 vm10.local ceph-mon[53712]: pgmap v2334: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:56.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:55 vm11.local ceph-mon[53973]: pgmap v2335: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:56.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:55 vm10.local ceph-mon[53712]: pgmap v2335: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:57.289 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:43:57.289 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:43:57.316 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:43:57.317 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:43:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:57 vm11.local ceph-mon[53973]: pgmap v2336: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:57 vm10.local ceph-mon[53712]: pgmap v2336: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:43:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:43:59 vm11.local ceph-mon[53973]: pgmap v2337: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:43:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:43:59 vm10.local ceph-mon[53712]: pgmap v2337: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:01 vm11.local ceph-mon[53973]: pgmap v2338: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:01 vm10.local ceph-mon[53712]: pgmap v2338: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:02.318 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:02.319 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:02.345 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:02.345 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:03 vm11.local ceph-mon[53973]: pgmap v2339: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:03 vm10.local ceph-mon[53712]: pgmap v2339: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:06.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:05 vm10.local ceph-mon[53712]: pgmap v2340: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:06.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:05 vm11.local ceph-mon[53973]: pgmap v2340: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:07 vm11.local ceph-mon[53973]: pgmap v2341: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:07.347 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:07.347 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:07.374 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:07.375 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:07 vm10.local ceph-mon[53712]: pgmap v2341: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:09 vm11.local ceph-mon[53973]: pgmap v2342: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:09 vm10.local ceph-mon[53712]: pgmap v2342: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:11 vm11.local ceph-mon[53973]: pgmap v2343: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:11 vm10.local ceph-mon[53712]: pgmap v2343: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:12.376 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:12.377 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:12.404 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:12.404 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:13 vm11.local ceph-mon[53973]: pgmap v2344: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:13 vm10.local ceph-mon[53712]: pgmap v2344: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:16.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:15 vm10.local ceph-mon[53712]: pgmap v2345: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:16.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:15 vm11.local ceph-mon[53973]: pgmap v2345: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:17 vm11.local ceph-mon[53973]: pgmap v2346: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:17.406 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:17.406 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:17.431 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:17.432 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:17 vm10.local ceph-mon[53712]: pgmap v2346: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:19 vm10.local ceph-mon[53712]: pgmap v2347: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:19 vm11.local ceph-mon[53973]: pgmap v2347: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:21 vm11.local ceph-mon[53973]: pgmap v2348: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:21 vm10.local ceph-mon[53712]: pgmap v2348: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:22.433 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:22.434 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:22.461 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:22.461 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:23 vm11.local ceph-mon[53973]: pgmap v2349: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:23 vm10.local ceph-mon[53712]: pgmap v2349: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:26.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:25 vm10.local ceph-mon[53712]: pgmap v2350: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:26.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:25 vm11.local ceph-mon[53973]: pgmap v2350: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:27 vm11.local ceph-mon[53973]: pgmap v2351: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:27 vm10.local ceph-mon[53712]: pgmap v2351: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:27.462 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:27.463 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:27.491 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:27.492 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:44:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:44:28.888 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:44:28.888 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:44:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:44:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:44:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:44:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:29 vm11.local ceph-mon[53973]: pgmap v2352: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:44:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:44:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:44:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:44:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:44:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:29 vm10.local ceph-mon[53712]: pgmap v2352: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:44:29.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:44:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:31 vm11.local ceph-mon[53973]: pgmap v2353: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:31 vm10.local ceph-mon[53712]: pgmap v2353: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:32.494 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:32.494 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:32.521 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:32.521 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:33 vm11.local ceph-mon[53973]: pgmap v2354: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:33 vm10.local ceph-mon[53712]: pgmap v2354: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:36.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:35 vm10.local ceph-mon[53712]: pgmap v2355: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:36.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:35 vm11.local ceph-mon[53973]: pgmap v2355: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:37 vm11.local ceph-mon[53973]: pgmap v2356: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:37 vm10.local ceph-mon[53712]: pgmap v2356: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:37.523 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:37.523 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:37.551 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:37.552 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:39 vm11.local ceph-mon[53973]: pgmap v2357: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:39 vm10.local ceph-mon[53712]: pgmap v2357: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:41 vm11.local ceph-mon[53973]: pgmap v2358: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:41 vm10.local ceph-mon[53712]: pgmap v2358: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:42.553 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:42.554 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:42.581 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:42.581 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:43 vm11.local ceph-mon[53973]: pgmap v2359: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:43 vm10.local ceph-mon[53712]: pgmap v2359: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:46.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:45 vm10.local ceph-mon[53712]: pgmap v2360: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:46.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:45 vm11.local ceph-mon[53973]: pgmap v2360: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:47 vm11.local ceph-mon[53973]: pgmap v2361: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:47 vm10.local ceph-mon[53712]: pgmap v2361: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:47.583 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:47.583 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:47.610 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:47.611 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:49 vm10.local ceph-mon[53712]: pgmap v2362: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:49 vm11.local ceph-mon[53973]: pgmap v2362: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:51 vm11.local ceph-mon[53973]: pgmap v2363: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:51 vm10.local ceph-mon[53712]: pgmap v2363: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:52.612 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:52.613 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:52.640 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:52.641 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:53 vm11.local ceph-mon[53973]: pgmap v2364: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:53 vm10.local ceph-mon[53712]: pgmap v2364: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:55 vm11.local ceph-mon[53973]: pgmap v2365: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:55 vm10.local ceph-mon[53712]: pgmap v2365: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:57 vm11.local ceph-mon[53973]: pgmap v2366: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:57 vm10.local ceph-mon[53712]: pgmap v2366: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:44:57.642 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:44:57.643 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:44:57.671 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:44:57.672 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:44:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:44:59 vm11.local ceph-mon[53973]: pgmap v2367: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:44:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:44:59 vm10.local ceph-mon[53712]: pgmap v2367: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:01 vm11.local ceph-mon[53973]: pgmap v2368: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:01 vm10.local ceph-mon[53712]: pgmap v2368: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:02.674 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:02.675 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:02.713 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:02.714 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:03 vm11.local ceph-mon[53973]: pgmap v2369: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:03 vm10.local ceph-mon[53712]: pgmap v2369: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:06.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:06 vm11.local ceph-mon[53973]: pgmap v2370: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:06.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:06 vm10.local ceph-mon[53712]: pgmap v2370: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:07 vm11.local ceph-mon[53973]: pgmap v2371: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:07.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:07 vm10.local ceph-mon[53712]: pgmap v2371: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:07.715 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:07.715 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:08.078 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:08.079 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:09 vm11.local ceph-mon[53973]: pgmap v2372: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:09 vm10.local ceph-mon[53712]: pgmap v2372: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:11.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:11 vm10.local ceph-mon[53712]: pgmap v2373: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:11 vm11.local ceph-mon[53973]: pgmap v2373: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:13.080 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:13.081 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:13.106 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:13.107 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:13 vm11.local ceph-mon[53973]: pgmap v2374: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:13 vm10.local ceph-mon[53712]: pgmap v2374: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:15 vm11.local ceph-mon[53973]: pgmap v2375: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:15 vm10.local ceph-mon[53712]: pgmap v2375: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:17.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:17 vm11.local ceph-mon[53973]: pgmap v2376: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:17.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:17 vm10.local ceph-mon[53712]: pgmap v2376: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:18.108 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:18.108 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:18.135 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:18.136 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:19 vm11.local ceph-mon[53973]: pgmap v2377: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:19.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:19 vm10.local ceph-mon[53712]: pgmap v2377: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:21 vm11.local ceph-mon[53973]: pgmap v2378: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:21.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:21 vm10.local ceph-mon[53712]: pgmap v2378: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:23.137 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:23.138 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:23.164 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:23.165 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:23 vm11.local ceph-mon[53973]: pgmap v2379: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:23 vm10.local ceph-mon[53712]: pgmap v2379: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:25 vm11.local ceph-mon[53973]: pgmap v2380: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:25 vm10.local ceph-mon[53712]: pgmap v2380: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:27 vm11.local ceph-mon[53973]: pgmap v2381: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:27 vm10.local ceph-mon[53712]: pgmap v2381: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:28.166 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:28.167 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:28.193 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:28.193 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:45:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:45:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:45:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:45:29.470 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:29 vm11.local ceph-mon[53973]: pgmap v2382: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:29.470 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:45:29.470 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:45:29.470 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:45:29.470 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:45:29.470 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:45:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:29 vm10.local ceph-mon[53712]: pgmap v2382: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:45:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:45:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:45:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:45:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:45:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:31 vm11.local ceph-mon[53973]: pgmap v2383: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:31 vm10.local ceph-mon[53712]: pgmap v2383: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:33.195 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:33.196 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:33.226 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:33.226 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:33 vm11.local ceph-mon[53973]: pgmap v2384: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:33 vm10.local ceph-mon[53712]: pgmap v2384: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:35 vm11.local ceph-mon[53973]: pgmap v2385: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:35 vm10.local ceph-mon[53712]: pgmap v2385: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:37 vm11.local ceph-mon[53973]: pgmap v2386: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:37 vm10.local ceph-mon[53712]: pgmap v2386: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:38.227 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:38.228 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:38.255 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:38.255 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:39 vm11.local ceph-mon[53973]: pgmap v2387: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:39 vm10.local ceph-mon[53712]: pgmap v2387: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:41 vm11.local ceph-mon[53973]: pgmap v2388: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:41 vm10.local ceph-mon[53712]: pgmap v2388: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:43.257 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:43.257 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:43.284 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:43.284 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:43 vm11.local ceph-mon[53973]: pgmap v2389: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:43 vm10.local ceph-mon[53712]: pgmap v2389: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:45 vm11.local ceph-mon[53973]: pgmap v2390: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:45 vm10.local ceph-mon[53712]: pgmap v2390: 97 pgs: 97 active+clean; 453 KiB data, 81 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:47 vm11.local ceph-mon[53973]: pgmap v2391: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:47 vm10.local ceph-mon[53712]: pgmap v2391: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:48.286 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:48.286 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:48.312 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:48.312 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:49.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:49 vm10.local ceph-mon[53712]: pgmap v2392: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:49 vm11.local ceph-mon[53973]: pgmap v2392: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:51 vm11.local ceph-mon[53973]: pgmap v2393: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:51.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:51 vm10.local ceph-mon[53712]: pgmap v2393: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:53.314 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:53.314 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:53 vm11.local ceph-mon[53973]: pgmap v2394: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:53.340 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:53.341 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:53 vm10.local ceph-mon[53712]: pgmap v2394: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:55 vm11.local ceph-mon[53973]: pgmap v2395: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:55 vm10.local ceph-mon[53712]: pgmap v2395: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:57 vm11.local ceph-mon[53973]: pgmap v2396: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:57 vm10.local ceph-mon[53712]: pgmap v2396: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:45:58.342 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:45:58.343 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:45:58.368 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:45:58.369 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:45:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:45:59 vm11.local ceph-mon[53973]: pgmap v2397: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:45:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:45:59 vm10.local ceph-mon[53712]: pgmap v2397: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:01 vm11.local ceph-mon[53973]: pgmap v2398: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:01.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:01 vm10.local ceph-mon[53712]: pgmap v2398: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:03 vm11.local ceph-mon[53973]: pgmap v2399: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:03.371 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:03.371 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:03.398 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:03.398 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:03 vm10.local ceph-mon[53712]: pgmap v2399: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:05 vm11.local ceph-mon[53973]: pgmap v2400: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:05 vm10.local ceph-mon[53712]: pgmap v2400: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:07 vm11.local ceph-mon[53973]: pgmap v2401: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:07 vm10.local ceph-mon[53712]: pgmap v2401: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:08.400 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:08.400 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:08.427 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:08.427 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:09 vm11.local ceph-mon[53973]: pgmap v2402: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:09 vm10.local ceph-mon[53712]: pgmap v2402: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:11 vm11.local ceph-mon[53973]: pgmap v2403: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:11.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:11 vm10.local ceph-mon[53712]: pgmap v2403: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:13 vm11.local ceph-mon[53973]: pgmap v2404: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:13.429 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:13.429 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:13 vm10.local ceph-mon[53712]: pgmap v2404: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:13.537 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:13.537 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:15 vm11.local ceph-mon[53973]: pgmap v2405: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:15 vm10.local ceph-mon[53712]: pgmap v2405: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:17 vm11.local ceph-mon[53973]: pgmap v2406: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:17 vm10.local ceph-mon[53712]: pgmap v2406: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:18.539 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:18.539 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:18.564 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:18.565 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:19 vm10.local ceph-mon[53712]: pgmap v2407: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:19 vm11.local ceph-mon[53973]: pgmap v2407: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:21 vm11.local ceph-mon[53973]: pgmap v2408: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:21 vm10.local ceph-mon[53712]: pgmap v2408: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:23 vm10.local ceph-mon[53712]: pgmap v2409: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:23.566 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:23.567 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:23 vm11.local ceph-mon[53973]: pgmap v2409: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:23.644 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:23.644 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:25 vm11.local ceph-mon[53973]: pgmap v2410: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:25 vm10.local ceph-mon[53712]: pgmap v2410: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:27 vm11.local ceph-mon[53973]: pgmap v2411: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:27.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:27 vm10.local ceph-mon[53712]: pgmap v2411: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:28.646 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:28.646 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:28.673 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:28.673 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:46:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:46:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:46:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:46:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:29 vm10.local ceph-mon[53712]: pgmap v2412: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:46:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:46:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:46:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:29 vm11.local ceph-mon[53973]: pgmap v2412: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:46:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:46:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:46:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:46:30.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:46:31.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:46:31.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:46:31.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:31 vm10.local ceph-mon[53712]: pgmap v2413: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:32.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:31 vm11.local ceph-mon[53973]: pgmap v2413: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:33 vm10.local ceph-mon[53712]: pgmap v2414: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:33 vm11.local ceph-mon[53973]: pgmap v2414: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:33.675 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:33.675 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:33.737 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:33.737 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:35 vm11.local ceph-mon[53973]: pgmap v2415: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:35 vm10.local ceph-mon[53712]: pgmap v2415: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:37 vm11.local ceph-mon[53973]: pgmap v2416: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:37 vm10.local ceph-mon[53712]: pgmap v2416: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:38.739 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:38.739 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:38.895 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:38.895 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:39 vm10.local ceph-mon[53712]: pgmap v2417: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:39 vm11.local ceph-mon[53973]: pgmap v2417: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:41 vm11.local ceph-mon[53973]: pgmap v2418: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:41 vm10.local ceph-mon[53712]: pgmap v2418: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:43 vm11.local ceph-mon[53973]: pgmap v2419: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:43 vm10.local ceph-mon[53712]: pgmap v2419: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:43.897 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:43.897 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:43.924 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:43.924 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:45 vm11.local ceph-mon[53973]: pgmap v2420: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:45 vm10.local ceph-mon[53712]: pgmap v2420: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:47 vm11.local ceph-mon[53973]: pgmap v2421: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:47 vm10.local ceph-mon[53712]: pgmap v2421: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:48.926 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:48.926 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:48.953 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:48.953 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:49 vm10.local ceph-mon[53712]: pgmap v2422: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:49 vm11.local ceph-mon[53973]: pgmap v2422: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:51 vm11.local ceph-mon[53973]: pgmap v2423: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:51 vm10.local ceph-mon[53712]: pgmap v2423: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:53 vm11.local ceph-mon[53973]: pgmap v2424: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:53 vm10.local ceph-mon[53712]: pgmap v2424: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:53.955 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:53.955 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:53.981 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:53.981 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:55 vm11.local ceph-mon[53973]: pgmap v2425: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:55 vm10.local ceph-mon[53712]: pgmap v2425: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:57 vm11.local ceph-mon[53973]: pgmap v2426: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:57.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:57 vm10.local ceph-mon[53712]: pgmap v2426: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:46:58.983 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:46:58.983 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:46:59.011 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:46:59.012 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:46:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:46:59 vm11.local ceph-mon[53973]: pgmap v2427: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:46:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:46:59 vm10.local ceph-mon[53712]: pgmap v2427: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:01 vm11.local ceph-mon[53973]: pgmap v2428: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:01 vm10.local ceph-mon[53712]: pgmap v2428: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:03 vm11.local ceph-mon[53973]: pgmap v2429: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:03 vm10.local ceph-mon[53712]: pgmap v2429: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:04.013 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:04.014 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:04.041 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:04.042 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:05 vm11.local ceph-mon[53973]: pgmap v2430: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:05 vm10.local ceph-mon[53712]: pgmap v2430: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:07 vm11.local ceph-mon[53973]: pgmap v2431: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:07 vm10.local ceph-mon[53712]: pgmap v2431: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:09.043 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:09.044 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:09.072 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:09.073 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:09 vm11.local ceph-mon[53973]: pgmap v2432: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:09 vm10.local ceph-mon[53712]: pgmap v2432: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:11 vm11.local ceph-mon[53973]: pgmap v2433: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:11 vm10.local ceph-mon[53712]: pgmap v2433: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:13 vm11.local ceph-mon[53973]: pgmap v2434: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:13 vm10.local ceph-mon[53712]: pgmap v2434: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:14.074 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:14.075 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:14.103 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:14.103 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:15 vm11.local ceph-mon[53973]: pgmap v2435: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:15 vm10.local ceph-mon[53712]: pgmap v2435: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:17 vm11.local ceph-mon[53973]: pgmap v2436: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:17.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:17 vm10.local ceph-mon[53712]: pgmap v2436: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:19.105 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:19.105 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:19.132 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:19.133 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:19.238 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:19 vm10.local ceph-mon[53712]: pgmap v2437: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:19 vm11.local ceph-mon[53973]: pgmap v2437: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:21 vm10.local ceph-mon[53712]: pgmap v2438: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:21 vm11.local ceph-mon[53973]: pgmap v2438: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:23 vm11.local ceph-mon[53973]: pgmap v2439: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:23 vm10.local ceph-mon[53712]: pgmap v2439: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:24.134 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:24.135 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:24.162 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:24.162 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:25 vm11.local ceph-mon[53973]: pgmap v2440: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:25 vm10.local ceph-mon[53712]: pgmap v2440: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:27 vm11.local ceph-mon[53973]: pgmap v2441: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:27.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:27 vm10.local ceph-mon[53712]: pgmap v2441: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:47:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:47:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:47:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:47:29.163 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:29.164 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:29.192 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:29.192 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:29 vm11.local ceph-mon[53973]: pgmap v2442: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:29.843 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:29 vm10.local ceph-mon[53712]: pgmap v2442: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:47:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:47:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:47:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:47:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:47:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:47:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:47:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:47:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:47:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:47:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:47:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:47:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:31 vm11.local ceph-mon[53973]: pgmap v2443: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:31.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:31 vm10.local ceph-mon[53712]: pgmap v2443: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:33 vm11.local ceph-mon[53973]: pgmap v2444: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:33 vm10.local ceph-mon[53712]: pgmap v2444: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:34.193 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:34.194 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:34.221 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:34.221 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:35 vm11.local ceph-mon[53973]: pgmap v2445: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:35 vm10.local ceph-mon[53712]: pgmap v2445: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:37 vm11.local ceph-mon[53973]: pgmap v2446: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:37 vm10.local ceph-mon[53712]: pgmap v2446: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:39.223 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:39.223 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:39.249 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:39.249 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:39 vm11.local ceph-mon[53973]: pgmap v2447: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:39 vm10.local ceph-mon[53712]: pgmap v2447: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:41 vm10.local ceph-mon[53712]: pgmap v2448: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:41 vm11.local ceph-mon[53973]: pgmap v2448: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:43 vm11.local ceph-mon[53973]: pgmap v2449: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:43 vm10.local ceph-mon[53712]: pgmap v2449: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:44.251 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:44.251 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:44.276 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:44.277 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:45 vm11.local ceph-mon[53973]: pgmap v2450: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:45 vm10.local ceph-mon[53712]: pgmap v2450: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:47 vm11.local ceph-mon[53973]: pgmap v2451: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:47 vm10.local ceph-mon[53712]: pgmap v2451: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:49 vm10.local ceph-mon[53712]: pgmap v2452: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:49.278 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:49.279 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:49.304 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:49.305 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:49 vm11.local ceph-mon[53973]: pgmap v2452: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:51 vm10.local ceph-mon[53712]: pgmap v2453: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:51 vm11.local ceph-mon[53973]: pgmap v2453: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:53 vm11.local ceph-mon[53973]: pgmap v2454: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:53 vm10.local ceph-mon[53712]: pgmap v2454: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:54.306 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:54.307 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:54.332 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:54.333 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:55 vm11.local ceph-mon[53973]: pgmap v2455: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:55 vm10.local ceph-mon[53712]: pgmap v2455: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:57 vm11.local ceph-mon[53973]: pgmap v2456: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:57.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:57 vm10.local ceph-mon[53712]: pgmap v2456: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:47:59.334 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:47:59.335 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:47:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:47:59 vm11.local ceph-mon[53973]: pgmap v2457: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:47:59.362 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:47:59.363 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:47:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:47:59 vm10.local ceph-mon[53712]: pgmap v2457: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:01 vm10.local ceph-mon[53712]: pgmap v2458: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:01 vm11.local ceph-mon[53973]: pgmap v2458: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:03 vm11.local ceph-mon[53973]: pgmap v2459: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:03 vm10.local ceph-mon[53712]: pgmap v2459: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:04.364 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:04.365 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:04.390 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:04.391 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:05 vm11.local ceph-mon[53973]: pgmap v2460: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:05 vm10.local ceph-mon[53712]: pgmap v2460: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:07.519 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:07 vm11.local ceph-mon[53973]: pgmap v2461: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:07 vm10.local ceph-mon[53712]: pgmap v2461: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:09 vm11.local ceph-mon[53973]: pgmap v2462: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:09.392 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:09.393 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:09.424 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:09.424 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:09 vm10.local ceph-mon[53712]: pgmap v2462: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:11 vm10.local ceph-mon[53712]: pgmap v2463: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:11 vm11.local ceph-mon[53973]: pgmap v2463: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:13 vm11.local ceph-mon[53973]: pgmap v2464: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:13 vm10.local ceph-mon[53712]: pgmap v2464: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:14.426 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:14.426 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:14.452 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:14.452 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:15 vm11.local ceph-mon[53973]: pgmap v2465: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:15 vm10.local ceph-mon[53712]: pgmap v2465: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:17 vm11.local ceph-mon[53973]: pgmap v2466: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:17 vm10.local ceph-mon[53712]: pgmap v2466: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:19.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:19 vm10.local ceph-mon[53712]: pgmap v2467: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:19 vm11.local ceph-mon[53973]: pgmap v2467: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:19.454 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:19.454 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:19.481 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:19.481 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:21 vm10.local ceph-mon[53712]: pgmap v2468: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:21 vm11.local ceph-mon[53973]: pgmap v2468: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:23 vm11.local ceph-mon[53973]: pgmap v2469: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:23 vm10.local ceph-mon[53712]: pgmap v2469: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:24.482 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:24.483 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:24.509 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:24.510 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:25 vm11.local ceph-mon[53973]: pgmap v2470: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:25 vm10.local ceph-mon[53712]: pgmap v2470: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:27 vm11.local ceph-mon[53973]: pgmap v2471: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:27 vm10.local ceph-mon[53712]: pgmap v2471: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:48:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:48:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:48:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:48:29.511 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:29.512 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:29.538 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:29.538 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:29 vm11.local ceph-mon[53973]: pgmap v2472: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:29 vm10.local ceph-mon[53712]: pgmap v2472: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:48:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:48:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:48:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:30 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:48:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:48:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:48:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:48:30.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:30 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:48:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:48:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:48:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:31 vm11.local ceph-mon[53973]: pgmap v2473: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:31.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:48:31.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:48:31.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:31 vm10.local ceph-mon[53712]: pgmap v2473: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:33 vm11.local ceph-mon[53973]: pgmap v2474: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:33 vm10.local ceph-mon[53712]: pgmap v2474: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:34.540 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:34.540 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:34.566 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:34.567 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:35 vm11.local ceph-mon[53973]: pgmap v2475: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:35 vm10.local ceph-mon[53712]: pgmap v2475: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:37 vm11.local ceph-mon[53973]: pgmap v2476: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:37 vm10.local ceph-mon[53712]: pgmap v2476: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:39 vm11.local ceph-mon[53973]: pgmap v2477: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:39 vm10.local ceph-mon[53712]: pgmap v2477: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:39.568 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:39.569 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:39.598 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:39.599 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:41 vm10.local ceph-mon[53712]: pgmap v2478: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:41 vm11.local ceph-mon[53973]: pgmap v2478: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:43 vm11.local ceph-mon[53973]: pgmap v2479: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:43.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:43 vm10.local ceph-mon[53712]: pgmap v2479: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:44.600 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:44.601 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:44.629 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:44.629 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:45 vm11.local ceph-mon[53973]: pgmap v2480: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:45 vm10.local ceph-mon[53712]: pgmap v2480: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:47 vm11.local ceph-mon[53973]: pgmap v2481: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:47 vm10.local ceph-mon[53712]: pgmap v2481: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:49 vm10.local ceph-mon[53712]: pgmap v2482: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:49 vm11.local ceph-mon[53973]: pgmap v2482: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:49.630 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:49.631 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:49.658 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:49.658 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:51 vm10.local ceph-mon[53712]: pgmap v2483: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:51 vm11.local ceph-mon[53973]: pgmap v2483: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:53 vm11.local ceph-mon[53973]: pgmap v2484: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:53 vm10.local ceph-mon[53712]: pgmap v2484: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:54.659 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:54.660 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:54.688 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:54.688 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:48:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:55 vm11.local ceph-mon[53973]: pgmap v2485: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:55 vm10.local ceph-mon[53712]: pgmap v2485: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:57 vm11.local ceph-mon[53973]: pgmap v2486: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:57 vm10.local ceph-mon[53712]: pgmap v2486: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:48:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:48:59 vm11.local ceph-mon[53973]: pgmap v2487: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:48:59 vm10.local ceph-mon[53712]: pgmap v2487: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:48:59.690 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:48:59.690 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:48:59.717 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:48:59.717 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:01 vm10.local ceph-mon[53712]: pgmap v2488: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:01 vm11.local ceph-mon[53973]: pgmap v2488: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:03 vm11.local ceph-mon[53973]: pgmap v2489: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:03.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:03 vm10.local ceph-mon[53712]: pgmap v2489: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:04.718 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:04.719 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:04.744 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:04.745 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:05 vm11.local ceph-mon[53973]: pgmap v2490: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:05 vm10.local ceph-mon[53712]: pgmap v2490: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:07 vm11.local ceph-mon[53973]: pgmap v2491: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:07 vm10.local ceph-mon[53712]: pgmap v2491: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:09 vm11.local ceph-mon[53973]: pgmap v2492: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:09 vm10.local ceph-mon[53712]: pgmap v2492: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:09.746 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:09.747 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:09.775 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:09.776 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:11 vm10.local ceph-mon[53712]: pgmap v2493: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:11 vm11.local ceph-mon[53973]: pgmap v2493: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:13 vm11.local ceph-mon[53973]: pgmap v2494: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:13 vm10.local ceph-mon[53712]: pgmap v2494: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:14.777 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:14.778 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:14.803 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:14.803 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:15 vm11.local ceph-mon[53973]: pgmap v2495: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:15 vm10.local ceph-mon[53712]: pgmap v2495: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:17 vm11.local ceph-mon[53973]: pgmap v2496: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:17 vm10.local ceph-mon[53712]: pgmap v2496: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:19 vm10.local ceph-mon[53712]: pgmap v2497: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:19 vm11.local ceph-mon[53973]: pgmap v2497: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:19.804 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:19.805 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:19.834 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:19.834 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:21 vm11.local ceph-mon[53973]: pgmap v2498: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:21 vm10.local ceph-mon[53712]: pgmap v2498: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:23 vm11.local ceph-mon[53973]: pgmap v2499: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:23 vm10.local ceph-mon[53712]: pgmap v2499: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:24.836 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:24.836 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:24.864 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:24.864 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:25 vm11.local ceph-mon[53973]: pgmap v2500: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:25 vm10.local ceph-mon[53712]: pgmap v2500: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:27 vm11.local ceph-mon[53973]: pgmap v2501: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:27 vm10.local ceph-mon[53712]: pgmap v2501: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:49:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:49:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:49:28.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:49:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:29 vm11.local ceph-mon[53973]: pgmap v2502: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:29.866 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:29.866 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:29.893 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:29.893 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:29 vm10.local ceph-mon[53712]: pgmap v2502: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:49:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:49:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:49:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:49:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:49:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:49:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:31 vm11.local ceph-mon[53973]: pgmap v2503: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:49:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:49:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:49:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:49:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:49:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:49:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:31 vm10.local ceph-mon[53712]: pgmap v2503: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:33 vm11.local ceph-mon[53973]: pgmap v2504: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:33 vm10.local ceph-mon[53712]: pgmap v2504: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:34.894 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:34.895 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:34.922 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:34.922 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:35 vm11.local ceph-mon[53973]: pgmap v2505: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:35 vm10.local ceph-mon[53712]: pgmap v2505: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:37 vm11.local ceph-mon[53973]: pgmap v2506: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:37 vm10.local ceph-mon[53712]: pgmap v2506: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:39 vm11.local ceph-mon[53973]: pgmap v2507: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:39 vm10.local ceph-mon[53712]: pgmap v2507: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:39.924 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:39.924 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:39.950 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:39.951 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:41 vm11.local ceph-mon[53973]: pgmap v2508: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:41 vm10.local ceph-mon[53712]: pgmap v2508: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:43 vm11.local ceph-mon[53973]: pgmap v2509: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:43 vm10.local ceph-mon[53712]: pgmap v2509: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:44.952 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:44.952 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:44.982 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:44.982 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:45 vm11.local ceph-mon[53973]: pgmap v2510: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:45 vm10.local ceph-mon[53712]: pgmap v2510: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:47 vm11.local ceph-mon[53973]: pgmap v2511: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:47 vm10.local ceph-mon[53712]: pgmap v2511: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:49.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:49 vm10.local ceph-mon[53712]: pgmap v2512: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:49 vm11.local ceph-mon[53973]: pgmap v2512: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:49.983 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:49.984 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:50.011 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:50.012 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:51 vm11.local ceph-mon[53973]: pgmap v2513: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:51 vm10.local ceph-mon[53712]: pgmap v2513: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:53 vm11.local ceph-mon[53973]: pgmap v2514: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:53 vm10.local ceph-mon[53712]: pgmap v2514: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:55.013 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:49:55.014 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:49:55.038 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:49:55.039 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:49:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:55 vm11.local ceph-mon[53973]: pgmap v2515: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:55 vm10.local ceph-mon[53712]: pgmap v2515: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:57 vm11.local ceph-mon[53973]: pgmap v2516: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:57 vm10.local ceph-mon[53712]: pgmap v2516: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:49:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:49:59 vm11.local ceph-mon[53973]: pgmap v2517: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:49:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:49:59 vm10.local ceph-mon[53712]: pgmap v2517: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:00.040 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:00.040 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:00.067 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:00.067 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:00 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T18:50:00.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:00 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T18:50:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:01 vm11.local ceph-mon[53973]: pgmap v2518: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:01 vm10.local ceph-mon[53712]: pgmap v2518: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:03 vm11.local ceph-mon[53973]: pgmap v2519: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:03 vm10.local ceph-mon[53712]: pgmap v2519: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:05.069 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:05.069 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:05.102 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:05.102 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:05 vm11.local ceph-mon[53973]: pgmap v2520: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:05 vm10.local ceph-mon[53712]: pgmap v2520: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:07 vm11.local ceph-mon[53973]: pgmap v2521: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:07 vm10.local ceph-mon[53712]: pgmap v2521: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:09 vm11.local ceph-mon[53973]: pgmap v2522: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:09 vm10.local ceph-mon[53712]: pgmap v2522: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:10.104 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:10.104 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:10.130 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:10.131 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:11 vm11.local ceph-mon[53973]: pgmap v2523: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:11 vm10.local ceph-mon[53712]: pgmap v2523: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:13 vm11.local ceph-mon[53973]: pgmap v2524: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:13.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:13 vm10.local ceph-mon[53712]: pgmap v2524: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:15.132 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:15.133 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:15.159 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:15.160 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:15 vm11.local ceph-mon[53973]: pgmap v2525: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:15 vm10.local ceph-mon[53712]: pgmap v2525: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:17 vm11.local ceph-mon[53973]: pgmap v2526: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:17 vm10.local ceph-mon[53712]: pgmap v2526: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:19.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:19 vm10.local ceph-mon[53712]: pgmap v2527: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:19 vm11.local ceph-mon[53973]: pgmap v2527: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:20.161 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:20.161 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:20.188 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:20.188 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:21 vm11.local ceph-mon[53973]: pgmap v2528: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:21 vm10.local ceph-mon[53712]: pgmap v2528: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:23 vm11.local ceph-mon[53973]: pgmap v2529: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:23 vm10.local ceph-mon[53712]: pgmap v2529: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:25.190 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:25.190 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:25.226 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:25.226 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:25 vm11.local ceph-mon[53973]: pgmap v2530: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:25 vm10.local ceph-mon[53712]: pgmap v2530: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:27 vm10.local ceph-mon[53712]: pgmap v2531: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:27 vm11.local ceph-mon[53973]: pgmap v2531: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:50:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:50:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:50:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:50:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:29 vm11.local ceph-mon[53973]: pgmap v2532: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:29 vm10.local ceph-mon[53712]: pgmap v2532: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:30.228 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:30.228 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:30.254 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:30.255 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:31.324 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:50:31.324 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:50:31.324 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:31 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:50:31.324 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:31 vm10.local ceph-mon[53712]: pgmap v2533: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:50:31.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:50:31.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:31 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:50:31.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:31 vm11.local ceph-mon[53973]: pgmap v2533: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:50:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:50:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:50:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:50:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:50:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:50:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:50:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:50:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:50:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:50:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:33 vm11.local ceph-mon[53973]: pgmap v2534: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:33 vm10.local ceph-mon[53712]: pgmap v2534: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:35.256 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:35.257 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:35.287 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:35.287 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:35 vm10.local ceph-mon[53712]: pgmap v2535: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:35 vm11.local ceph-mon[53973]: pgmap v2535: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:37.392 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:37 vm11.local ceph-mon[53973]: pgmap v2536: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:37 vm10.local ceph-mon[53712]: pgmap v2536: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:39 vm10.local ceph-mon[53712]: pgmap v2537: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:39 vm11.local ceph-mon[53973]: pgmap v2537: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:40.289 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:40.289 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:40.315 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:40.315 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:41 vm11.local ceph-mon[53973]: pgmap v2538: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:41 vm10.local ceph-mon[53712]: pgmap v2538: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:43.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:43 vm10.local ceph-mon[53712]: pgmap v2539: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:43 vm11.local ceph-mon[53973]: pgmap v2539: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:45.316 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:45.317 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:45.344 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:45.344 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:45 vm10.local ceph-mon[53712]: pgmap v2540: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:45 vm11.local ceph-mon[53973]: pgmap v2540: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:47 vm10.local ceph-mon[53712]: pgmap v2541: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:47 vm11.local ceph-mon[53973]: pgmap v2541: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:49 vm10.local ceph-mon[53712]: pgmap v2542: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:49 vm11.local ceph-mon[53973]: pgmap v2542: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:50.346 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:50.347 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:50.374 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:50.374 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:51 vm11.local ceph-mon[53973]: pgmap v2543: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:51 vm10.local ceph-mon[53712]: pgmap v2543: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:53 vm10.local ceph-mon[53712]: pgmap v2544: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:53 vm11.local ceph-mon[53973]: pgmap v2544: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:55.375 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:50:55.376 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:50:55.401 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:50:55.402 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:50:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:55 vm10.local ceph-mon[53712]: pgmap v2545: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:55 vm11.local ceph-mon[53973]: pgmap v2545: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:57 vm10.local ceph-mon[53712]: pgmap v2546: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:57 vm11.local ceph-mon[53973]: pgmap v2546: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:50:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:50:59 vm10.local ceph-mon[53712]: pgmap v2547: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:50:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:50:59 vm11.local ceph-mon[53973]: pgmap v2547: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:00.403 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:00.404 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:00.429 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:00.430 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:01 vm11.local ceph-mon[53973]: pgmap v2548: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:01 vm10.local ceph-mon[53712]: pgmap v2548: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:03.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:03 vm10.local ceph-mon[53712]: pgmap v2549: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:03 vm11.local ceph-mon[53973]: pgmap v2549: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:05.431 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:05.432 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:05 vm10.local ceph-mon[53712]: pgmap v2550: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:05.459 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:05.459 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:05 vm11.local ceph-mon[53973]: pgmap v2550: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:07.392 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:07 vm11.local ceph-mon[53973]: pgmap v2551: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:07 vm10.local ceph-mon[53712]: pgmap v2551: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:09 vm10.local ceph-mon[53712]: pgmap v2552: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:09 vm11.local ceph-mon[53973]: pgmap v2552: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:10.460 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:10.461 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:10.487 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:10.487 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:11 vm11.local ceph-mon[53973]: pgmap v2553: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:11 vm10.local ceph-mon[53712]: pgmap v2553: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:13 vm10.local ceph-mon[53712]: pgmap v2554: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:13 vm11.local ceph-mon[53973]: pgmap v2554: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:15 vm10.local ceph-mon[53712]: pgmap v2555: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:15.489 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:15.489 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:15.515 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:15.515 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:15 vm11.local ceph-mon[53973]: pgmap v2555: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:17 vm10.local ceph-mon[53712]: pgmap v2556: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:17 vm11.local ceph-mon[53973]: pgmap v2556: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:19 vm10.local ceph-mon[53712]: pgmap v2557: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:19 vm11.local ceph-mon[53973]: pgmap v2557: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:20.517 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:20.517 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:20.545 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:20.545 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:21 vm11.local ceph-mon[53973]: pgmap v2558: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:21 vm10.local ceph-mon[53712]: pgmap v2558: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:23 vm10.local ceph-mon[53712]: pgmap v2559: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:23 vm11.local ceph-mon[53973]: pgmap v2559: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:25 vm10.local ceph-mon[53712]: pgmap v2560: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:25.547 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:25.547 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:25.572 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:25.573 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:25 vm11.local ceph-mon[53973]: pgmap v2560: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:27 vm10.local ceph-mon[53712]: pgmap v2561: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:27 vm11.local ceph-mon[53973]: pgmap v2561: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:51:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:51:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:51:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:51:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:29 vm11.local ceph-mon[53973]: pgmap v2562: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:29 vm10.local ceph-mon[53712]: pgmap v2562: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:30.574 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:30.574 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:30.600 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:30.601 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:31 vm11.local ceph-mon[53973]: pgmap v2563: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:31 vm10.local ceph-mon[53712]: pgmap v2563: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:51:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:51:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:51:32.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:51:32.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:51:32.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:51:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:51:33.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:51:33.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:33 vm10.local ceph-mon[53712]: pgmap v2564: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:51:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:51:33.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:33 vm11.local ceph-mon[53973]: pgmap v2564: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:35 vm10.local ceph-mon[53712]: pgmap v2565: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:35 vm11.local ceph-mon[53973]: pgmap v2565: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:35.602 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:35.602 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:35.628 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:35.629 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:37.393 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:37 vm11.local ceph-mon[53973]: pgmap v2566: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:37 vm10.local ceph-mon[53712]: pgmap v2566: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:39 vm10.local ceph-mon[53712]: pgmap v2567: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:39 vm11.local ceph-mon[53973]: pgmap v2567: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:40.630 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:40.630 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:40.656 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:40.657 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:41 vm11.local ceph-mon[53973]: pgmap v2568: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:41 vm10.local ceph-mon[53712]: pgmap v2568: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:43 vm10.local ceph-mon[53712]: pgmap v2569: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:43.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:43 vm11.local ceph-mon[53973]: pgmap v2569: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:45 vm10.local ceph-mon[53712]: pgmap v2570: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:45 vm11.local ceph-mon[53973]: pgmap v2570: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:45.658 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:45.659 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:45.685 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:45.685 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:47 vm10.local ceph-mon[53712]: pgmap v2571: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:47 vm11.local ceph-mon[53973]: pgmap v2571: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:49 vm10.local ceph-mon[53712]: pgmap v2572: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:49 vm11.local ceph-mon[53973]: pgmap v2572: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:50.687 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:50.688 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:50.714 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:50.715 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:51 vm11.local ceph-mon[53973]: pgmap v2573: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:51 vm10.local ceph-mon[53712]: pgmap v2573: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:53 vm10.local ceph-mon[53712]: pgmap v2574: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:53 vm11.local ceph-mon[53973]: pgmap v2574: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:55 vm10.local ceph-mon[53712]: pgmap v2575: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:55 vm11.local ceph-mon[53973]: pgmap v2575: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:55.716 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:51:55.717 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:51:55.743 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:51:55.744 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:51:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:57 vm10.local ceph-mon[53712]: pgmap v2576: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:57 vm11.local ceph-mon[53973]: pgmap v2576: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:51:59.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:51:59 vm10.local ceph-mon[53712]: pgmap v2577: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:51:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:51:59 vm11.local ceph-mon[53973]: pgmap v2577: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:00.745 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:00.746 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:00.771 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:00.772 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:01 vm11.local ceph-mon[53973]: pgmap v2578: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:01.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:01 vm10.local ceph-mon[53712]: pgmap v2578: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:03 vm10.local ceph-mon[53712]: pgmap v2579: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:03 vm11.local ceph-mon[53973]: pgmap v2579: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:05 vm10.local ceph-mon[53712]: pgmap v2580: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:05 vm11.local ceph-mon[53973]: pgmap v2580: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:05.773 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:05.774 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:05.800 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:05.800 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:07.393 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:07 vm11.local ceph-mon[53973]: pgmap v2581: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:07 vm10.local ceph-mon[53712]: pgmap v2581: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:09 vm10.local ceph-mon[53712]: pgmap v2582: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:09 vm11.local ceph-mon[53973]: pgmap v2582: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:10.801 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:10.802 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:10.828 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:10.829 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:11 vm11.local ceph-mon[53973]: pgmap v2583: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:11 vm10.local ceph-mon[53712]: pgmap v2583: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:13 vm10.local ceph-mon[53712]: pgmap v2584: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:13 vm11.local ceph-mon[53973]: pgmap v2584: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:15 vm10.local ceph-mon[53712]: pgmap v2585: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:15 vm11.local ceph-mon[53973]: pgmap v2585: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:15.830 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:15.831 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:15.860 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:15.861 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:17 vm10.local ceph-mon[53712]: pgmap v2586: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:17 vm11.local ceph-mon[53973]: pgmap v2586: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:19 vm10.local ceph-mon[53712]: pgmap v2587: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:19 vm11.local ceph-mon[53973]: pgmap v2587: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:20.862 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:20.863 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:20.891 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:20.892 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:21 vm11.local ceph-mon[53973]: pgmap v2588: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:21.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:21 vm10.local ceph-mon[53712]: pgmap v2588: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:23 vm10.local ceph-mon[53712]: pgmap v2589: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:23 vm11.local ceph-mon[53973]: pgmap v2589: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:25 vm10.local ceph-mon[53712]: pgmap v2590: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:25 vm11.local ceph-mon[53973]: pgmap v2590: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:25.893 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:25.893 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:25.920 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:25.921 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:27 vm10.local ceph-mon[53712]: pgmap v2591: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:27 vm11.local ceph-mon[53973]: pgmap v2591: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:52:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:52:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:52:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:52:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:29 vm11.local ceph-mon[53973]: pgmap v2592: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:29 vm10.local ceph-mon[53712]: pgmap v2592: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:30.922 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:30.923 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:30.951 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:30.951 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:31 vm11.local ceph-mon[53973]: pgmap v2593: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:31 vm10.local ceph-mon[53712]: pgmap v2593: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:32.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:52:32.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:52:32.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:52:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:52:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:52:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:52:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:52:33.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:52:33.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:33 vm10.local ceph-mon[53712]: pgmap v2594: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:52:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:52:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:33 vm11.local ceph-mon[53973]: pgmap v2594: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:35 vm10.local ceph-mon[53712]: pgmap v2595: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:35 vm11.local ceph-mon[53973]: pgmap v2595: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:35.953 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:35.953 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:35.981 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:35.982 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:37.393 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:37 vm11.local ceph-mon[53973]: pgmap v2596: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:37 vm10.local ceph-mon[53712]: pgmap v2596: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:39 vm10.local ceph-mon[53712]: pgmap v2597: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:39 vm11.local ceph-mon[53973]: pgmap v2597: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:40.983 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:40.984 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:41.012 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:41.013 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:41 vm11.local ceph-mon[53973]: pgmap v2598: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:41 vm10.local ceph-mon[53712]: pgmap v2598: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:43 vm10.local ceph-mon[53712]: pgmap v2599: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:43 vm11.local ceph-mon[53973]: pgmap v2599: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:45 vm10.local ceph-mon[53712]: pgmap v2600: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:45 vm11.local ceph-mon[53973]: pgmap v2600: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:46.014 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:46.014 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:46.040 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:46.040 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:47 vm10.local ceph-mon[53712]: pgmap v2601: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:47 vm11.local ceph-mon[53973]: pgmap v2601: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:49.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:49 vm10.local ceph-mon[53712]: pgmap v2602: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:49 vm11.local ceph-mon[53973]: pgmap v2602: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:51.042 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:51.042 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:51.067 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:51.068 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:51 vm11.local ceph-mon[53973]: pgmap v2603: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:51 vm10.local ceph-mon[53712]: pgmap v2603: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:53 vm11.local ceph-mon[53973]: pgmap v2604: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:53.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:53 vm10.local ceph-mon[53712]: pgmap v2604: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:55 vm10.local ceph-mon[53712]: pgmap v2605: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:55 vm11.local ceph-mon[53973]: pgmap v2605: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:56.069 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:52:56.070 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:52:56.130 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:52:56.131 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:52:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:57 vm11.local ceph-mon[53973]: pgmap v2606: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:57 vm10.local ceph-mon[53712]: pgmap v2606: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:52:59.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:52:59 vm10.local ceph-mon[53712]: pgmap v2607: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:52:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:52:59 vm11.local ceph-mon[53973]: pgmap v2607: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:01.133 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:01.133 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:01.236 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:01.237 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:01 vm11.local ceph-mon[53973]: pgmap v2608: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:01.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:01 vm10.local ceph-mon[53712]: pgmap v2608: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:03 vm11.local ceph-mon[53973]: pgmap v2609: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:03.686 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:03 vm10.local ceph-mon[53712]: pgmap v2609: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:05 vm11.local ceph-mon[53973]: pgmap v2610: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:05.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:05 vm10.local ceph-mon[53712]: pgmap v2610: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:06.239 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:06.239 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:06.334 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:06.335 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:07 vm11.local ceph-mon[53973]: pgmap v2611: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:07.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:07 vm10.local ceph-mon[53712]: pgmap v2611: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:09 vm10.local ceph-mon[53712]: pgmap v2612: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:09 vm11.local ceph-mon[53973]: pgmap v2612: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:11.336 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:11.336 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:11.399 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:11.399 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:11.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:11 vm11.local ceph-mon[53973]: pgmap v2613: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:11.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:11 vm10.local ceph-mon[53712]: pgmap v2613: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:13 vm10.local ceph-mon[53712]: pgmap v2614: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:13 vm11.local ceph-mon[53973]: pgmap v2614: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:15 vm11.local ceph-mon[53973]: pgmap v2615: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:15.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:15 vm10.local ceph-mon[53712]: pgmap v2615: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:16.401 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:16.401 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:16.497 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:16.498 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:17 vm10.local ceph-mon[53712]: pgmap v2616: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:17 vm11.local ceph-mon[53973]: pgmap v2616: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:19 vm11.local ceph-mon[53973]: pgmap v2617: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:53:19.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:19 vm10.local ceph-mon[53712]: pgmap v2617: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:53:21.499 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:21.500 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:21.708 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:21.708 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:21.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:21 vm11.local ceph-mon[53973]: pgmap v2618: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:53:21.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:21 vm10.local ceph-mon[53712]: pgmap v2618: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:53:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:23 vm10.local ceph-mon[53712]: pgmap v2619: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:23 vm11.local ceph-mon[53973]: pgmap v2619: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:25 vm10.local ceph-mon[53712]: pgmap v2620: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:53:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:25 vm11.local ceph-mon[53973]: pgmap v2620: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T18:53:26.710 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:26.710 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:26.738 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:26.739 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:27 vm10.local ceph-mon[53712]: pgmap v2621: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:27 vm11.local ceph-mon[53973]: pgmap v2621: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:53:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:53:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:53:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:53:29.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:29 vm11.local ceph-mon[53973]: pgmap v2622: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:29.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:29 vm10.local ceph-mon[53712]: pgmap v2622: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:31 vm10.local ceph-mon[53712]: pgmap v2623: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:31.740 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:31.741 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:31.775 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:31.775 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:31.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:31 vm11.local ceph-mon[53973]: pgmap v2623: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:53:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:53:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:53:33.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:53:33.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:53:33.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:53:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:53:34.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:53:34.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:33 vm10.local ceph-mon[53712]: pgmap v2624: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:53:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:53:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:33 vm11.local ceph-mon[53973]: pgmap v2624: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:35 vm10.local ceph-mon[53712]: pgmap v2625: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:35 vm11.local ceph-mon[53973]: pgmap v2625: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:36.777 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:36.777 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:36.807 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:36.807 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:37.393 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:37 vm11.local ceph-mon[53973]: pgmap v2626: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:37 vm10.local ceph-mon[53712]: pgmap v2626: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:39 vm10.local ceph-mon[53712]: pgmap v2627: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:39 vm11.local ceph-mon[53973]: pgmap v2627: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:41.809 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:41.809 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:41.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:41 vm10.local ceph-mon[53712]: pgmap v2628: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:41.936 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:41.938 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:42.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:41 vm11.local ceph-mon[53973]: pgmap v2628: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:43 vm11.local ceph-mon[53973]: pgmap v2629: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:43.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:43 vm10.local ceph-mon[53712]: pgmap v2629: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:45 vm11.local ceph-mon[53973]: pgmap v2630: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:45 vm10.local ceph-mon[53712]: pgmap v2630: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:46.940 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:46.941 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:47.058 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:47.059 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:47.662 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:47 vm11.local ceph-mon[53973]: pgmap v2631: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:47 vm10.local ceph-mon[53712]: pgmap v2631: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:49.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:49 vm11.local ceph-mon[53973]: pgmap v2632: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:49.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:49 vm10.local ceph-mon[53712]: pgmap v2632: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:51.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:51 vm10.local ceph-mon[53712]: pgmap v2633: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:52.060 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:52.060 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:52.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:51 vm11.local ceph-mon[53973]: pgmap v2633: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:52.094 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:52.095 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:53 vm11.local ceph-mon[53973]: pgmap v2634: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:53 vm10.local ceph-mon[53712]: pgmap v2634: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:55 vm11.local ceph-mon[53973]: pgmap v2635: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:55.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:55 vm10.local ceph-mon[53712]: pgmap v2635: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:57.096 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:53:57.097 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:53:57.156 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:53:57.157 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:53:57.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:57 vm10.local ceph-mon[53712]: pgmap v2636: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:57 vm11.local ceph-mon[53973]: pgmap v2636: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:53:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:53:59 vm11.local ceph-mon[53973]: pgmap v2637: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:53:59.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:53:59 vm10.local ceph-mon[53712]: pgmap v2637: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:02.091 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:01 vm11.local ceph-mon[53973]: pgmap v2638: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:02.159 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:02.159 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:02.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:01 vm10.local ceph-mon[53712]: pgmap v2638: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:02.289 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:02.290 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:03 vm11.local ceph-mon[53973]: pgmap v2639: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:03 vm10.local ceph-mon[53712]: pgmap v2639: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:05 vm11.local ceph-mon[53973]: pgmap v2640: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:05 vm10.local ceph-mon[53712]: pgmap v2640: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:07.292 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:07.292 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:07.413 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:07.414 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:08.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:08 vm11.local ceph-mon[53973]: pgmap v2641: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:08.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:08 vm10.local ceph-mon[53712]: pgmap v2641: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:10.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:10 vm11.local ceph-mon[53973]: pgmap v2642: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:10.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:10 vm10.local ceph-mon[53712]: pgmap v2642: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:12.415 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:12.416 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:12.444 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:12.444 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:12 vm10.local ceph-mon[53712]: pgmap v2643: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:12 vm11.local ceph-mon[53973]: pgmap v2643: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:14 vm10.local ceph-mon[53712]: pgmap v2644: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:14 vm11.local ceph-mon[53973]: pgmap v2644: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:16 vm10.local ceph-mon[53712]: pgmap v2645: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:16 vm11.local ceph-mon[53973]: pgmap v2645: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:17.446 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:17.446 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:17.474 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:17.475 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:18 vm10.local ceph-mon[53712]: pgmap v2646: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:18 vm11.local ceph-mon[53973]: pgmap v2646: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:20 vm10.local ceph-mon[53712]: pgmap v2647: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:20 vm11.local ceph-mon[53973]: pgmap v2647: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:22.477 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:22.477 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:22.505 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:22.505 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:22 vm10.local ceph-mon[53712]: pgmap v2648: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:22 vm11.local ceph-mon[53973]: pgmap v2648: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:24 vm10.local ceph-mon[53712]: pgmap v2649: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:24 vm11.local ceph-mon[53973]: pgmap v2649: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:26 vm10.local ceph-mon[53712]: pgmap v2650: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:26 vm11.local ceph-mon[53973]: pgmap v2650: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:27.507 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:27.507 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:27.534 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:27.535 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:28 vm10.local ceph-mon[53712]: pgmap v2651: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:54:29.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:54:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:28 vm11.local ceph-mon[53973]: pgmap v2651: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:54:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:54:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:30 vm11.local ceph-mon[53973]: pgmap v2652: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:30 vm10.local ceph-mon[53712]: pgmap v2652: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:32.536 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:32.537 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:32.564 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:32.565 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:33.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:32 vm10.local ceph-mon[53712]: pgmap v2653: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:33.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:54:33.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:54:33.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:32 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:54:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:32 vm11.local ceph-mon[53973]: pgmap v2653: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:54:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:54:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:32 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:54:34.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:34 vm10.local ceph-mon[53712]: pgmap v2654: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:34.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:54:34.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:54:34.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:34 vm11.local ceph-mon[53973]: pgmap v2654: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:34.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:54:34.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:54:36.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:36 vm10.local ceph-mon[53712]: pgmap v2655: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:37.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:36 vm11.local ceph-mon[53973]: pgmap v2655: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:37.566 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:37.567 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:37.593 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:37.593 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:38 vm10.local ceph-mon[53712]: pgmap v2656: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:38.991 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:38 vm11.local ceph-mon[53973]: pgmap v2656: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:40 vm10.local ceph-mon[53712]: pgmap v2657: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:40 vm11.local ceph-mon[53973]: pgmap v2657: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:42.595 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:42.595 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:42.622 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:42.623 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:42.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:42 vm10.local ceph-mon[53712]: pgmap v2658: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:43.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:42 vm11.local ceph-mon[53973]: pgmap v2658: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:44.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:44 vm10.local ceph-mon[53712]: pgmap v2659: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:44 vm11.local ceph-mon[53973]: pgmap v2659: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:46.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:46 vm10.local ceph-mon[53712]: pgmap v2660: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:46 vm11.local ceph-mon[53973]: pgmap v2660: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:47.624 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:47.625 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:47.652 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:47.652 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:48 vm10.local ceph-mon[53712]: pgmap v2661: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:48 vm11.local ceph-mon[53973]: pgmap v2661: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:50.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:50 vm10.local ceph-mon[53712]: pgmap v2662: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:51.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:50 vm11.local ceph-mon[53973]: pgmap v2662: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:52.654 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:52.654 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:52.681 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:52.681 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:52 vm11.local ceph-mon[53973]: pgmap v2663: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:52 vm10.local ceph-mon[53712]: pgmap v2663: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:54 vm11.local ceph-mon[53973]: pgmap v2664: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:54 vm10.local ceph-mon[53712]: pgmap v2664: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:56 vm11.local ceph-mon[53973]: pgmap v2665: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:56 vm10.local ceph-mon[53712]: pgmap v2665: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:54:57.683 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:54:57.683 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:54:57.710 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:54:57.710 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:54:59.061 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:54:58 vm10.local ceph-mon[53712]: pgmap v2666: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:54:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:54:58 vm11.local ceph-mon[53973]: pgmap v2666: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:00 vm11.local ceph-mon[53973]: pgmap v2667: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:00 vm10.local ceph-mon[53712]: pgmap v2667: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:02.712 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:02.712 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:02.739 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:02.739 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:02 vm11.local ceph-mon[53973]: pgmap v2668: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:02 vm10.local ceph-mon[53712]: pgmap v2668: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:04 vm11.local ceph-mon[53973]: pgmap v2669: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:04 vm10.local ceph-mon[53712]: pgmap v2669: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:06 vm11.local ceph-mon[53973]: pgmap v2670: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:06 vm10.local ceph-mon[53712]: pgmap v2670: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:07.740 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:07.741 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:07.769 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:07.769 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:09.062 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:08 vm10.local ceph-mon[53712]: pgmap v2671: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:08 vm11.local ceph-mon[53973]: pgmap v2671: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:10 vm11.local ceph-mon[53973]: pgmap v2672: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:10 vm10.local ceph-mon[53712]: pgmap v2672: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:12.771 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:12.771 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:12.800 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:12.801 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:12 vm11.local ceph-mon[53973]: pgmap v2673: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:12 vm10.local ceph-mon[53712]: pgmap v2673: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:14 vm11.local ceph-mon[53973]: pgmap v2674: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:14 vm10.local ceph-mon[53712]: pgmap v2674: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:17.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:16 vm11.local ceph-mon[53973]: pgmap v2675: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:16 vm10.local ceph-mon[53712]: pgmap v2675: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:17.802 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:17.803 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:17.828 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:17.829 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:19.063 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:18 vm10.local ceph-mon[53712]: pgmap v2676: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:18 vm11.local ceph-mon[53973]: pgmap v2676: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:20 vm11.local ceph-mon[53973]: pgmap v2677: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:20 vm10.local ceph-mon[53712]: pgmap v2677: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:22.830 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:22.831 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:22.858 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:22.859 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:22 vm11.local ceph-mon[53973]: pgmap v2678: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:22 vm10.local ceph-mon[53712]: pgmap v2678: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:24 vm10.local ceph-mon[53712]: pgmap v2679: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:24 vm11.local ceph-mon[53973]: pgmap v2679: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:27.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:26 vm10.local ceph-mon[53712]: pgmap v2680: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:26 vm11.local ceph-mon[53973]: pgmap v2680: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:27.861 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:27.862 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:28.003 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:28.005 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:29 vm11.local ceph-mon[53973]: pgmap v2681: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:55:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:55:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:29 vm10.local ceph-mon[53712]: pgmap v2681: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:55:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:55:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:31 vm10.local ceph-mon[53712]: pgmap v2682: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:31 vm11.local ceph-mon[53973]: pgmap v2682: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:33.006 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:33.006 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:33.033 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:33.033 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:33.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:32 vm11.local ceph-mon[53973]: pgmap v2683: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:33.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:32 vm10.local ceph-mon[53712]: pgmap v2683: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:33.971 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:55:33.971 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:55:33.971 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:55:33.971 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:55:33.971 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:55:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:55:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:55:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:55:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:55:34.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:55:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:34 vm11.local ceph-mon[53973]: pgmap v2684: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:34 vm10.local ceph-mon[53712]: pgmap v2684: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:37.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:36 vm11.local ceph-mon[53973]: pgmap v2685: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:37.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:36 vm10.local ceph-mon[53712]: pgmap v2685: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:38.035 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:38.035 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:38.063 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:38.063 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:38.991 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:38 vm11.local ceph-mon[53973]: pgmap v2686: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:39.066 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:38 vm10.local ceph-mon[53712]: pgmap v2686: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:40 vm11.local ceph-mon[53973]: pgmap v2687: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:40 vm10.local ceph-mon[53712]: pgmap v2687: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:43.065 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:43.065 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:43.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:42 vm11.local ceph-mon[53973]: pgmap v2688: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:43.092 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:43.092 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:42 vm10.local ceph-mon[53712]: pgmap v2688: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:44 vm11.local ceph-mon[53973]: pgmap v2689: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:44 vm10.local ceph-mon[53712]: pgmap v2689: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:46 vm11.local ceph-mon[53973]: pgmap v2690: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:46 vm10.local ceph-mon[53712]: pgmap v2690: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:48.094 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:48.094 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:48.119 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:48.120 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:49.067 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:48 vm10.local ceph-mon[53712]: pgmap v2691: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:48 vm11.local ceph-mon[53973]: pgmap v2691: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:51.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:50 vm11.local ceph-mon[53973]: pgmap v2692: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:50 vm10.local ceph-mon[53712]: pgmap v2692: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:52 vm11.local ceph-mon[53973]: pgmap v2693: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:53.121 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:53.122 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:53.148 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:53.149 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:52 vm10.local ceph-mon[53712]: pgmap v2693: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:54 vm11.local ceph-mon[53973]: pgmap v2694: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:54 vm10.local ceph-mon[53712]: pgmap v2694: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:56 vm11.local ceph-mon[53973]: pgmap v2695: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:56 vm10.local ceph-mon[53712]: pgmap v2695: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:55:58.151 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:55:58.151 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:55:58.186 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:55:58.187 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:55:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:55:58 vm11.local ceph-mon[53973]: pgmap v2696: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:55:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:55:58 vm10.local ceph-mon[53712]: pgmap v2696: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:00 vm11.local ceph-mon[53973]: pgmap v2697: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:00 vm10.local ceph-mon[53712]: pgmap v2697: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:02 vm11.local ceph-mon[53973]: pgmap v2698: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:02 vm10.local ceph-mon[53712]: pgmap v2698: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:03.188 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:03.188 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:03.214 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:03.215 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:04 vm10.local ceph-mon[53712]: pgmap v2699: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:04 vm11.local ceph-mon[53973]: pgmap v2699: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:06 vm11.local ceph-mon[53973]: pgmap v2700: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:06 vm10.local ceph-mon[53712]: pgmap v2700: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:08.216 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:08.217 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:08.245 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:08.245 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:08 vm11.local ceph-mon[53973]: pgmap v2701: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:08 vm10.local ceph-mon[53712]: pgmap v2701: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:10 vm11.local ceph-mon[53973]: pgmap v2702: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:10 vm10.local ceph-mon[53712]: pgmap v2702: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:13.247 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:13.247 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:13.275 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:13.275 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:12 vm11.local ceph-mon[53973]: pgmap v2703: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:12 vm10.local ceph-mon[53712]: pgmap v2703: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:14 vm11.local ceph-mon[53973]: pgmap v2704: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:14 vm10.local ceph-mon[53712]: pgmap v2704: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:16 vm11.local ceph-mon[53973]: pgmap v2705: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:16 vm10.local ceph-mon[53712]: pgmap v2705: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:18.277 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:18.277 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:18.303 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:18.304 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:19.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:18 vm10.local ceph-mon[53712]: pgmap v2706: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:18 vm11.local ceph-mon[53973]: pgmap v2706: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:20 vm11.local ceph-mon[53973]: pgmap v2707: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:20 vm10.local ceph-mon[53712]: pgmap v2707: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:23.306 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:23.306 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:23.335 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:23.335 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:22 vm11.local ceph-mon[53973]: pgmap v2708: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:22 vm10.local ceph-mon[53712]: pgmap v2708: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:24 vm11.local ceph-mon[53973]: pgmap v2709: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:24 vm10.local ceph-mon[53712]: pgmap v2709: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:26 vm11.local ceph-mon[53973]: pgmap v2710: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:26 vm10.local ceph-mon[53712]: pgmap v2710: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:28.337 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:28.337 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:28.364 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:28.364 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:28 vm11.local ceph-mon[53973]: pgmap v2711: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:56:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:56:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:28 vm10.local ceph-mon[53712]: pgmap v2711: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:56:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:56:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:31 vm11.local ceph-mon[53973]: pgmap v2712: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:31.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:31 vm10.local ceph-mon[53712]: pgmap v2712: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:33 vm11.local ceph-mon[53973]: pgmap v2713: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:33.366 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:33.367 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:33.392 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:33.393 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:33.429 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:33 vm10.local ceph-mon[53712]: pgmap v2713: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:56:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:56:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:56:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:56:34.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:56:34.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:56:34.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:56:34.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:56:34.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:56:34.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:56:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:35 vm11.local ceph-mon[53973]: pgmap v2714: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:35 vm10.local ceph-mon[53712]: pgmap v2714: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:37 vm11.local ceph-mon[53973]: pgmap v2715: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:37 vm10.local ceph-mon[53712]: pgmap v2715: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:38.394 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:38.395 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:38.420 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:38.421 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:39 vm11.local ceph-mon[53973]: pgmap v2716: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:39 vm10.local ceph-mon[53712]: pgmap v2716: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:41 vm11.local ceph-mon[53973]: pgmap v2717: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:41.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:41 vm10.local ceph-mon[53712]: pgmap v2717: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:43 vm11.local ceph-mon[53973]: pgmap v2718: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:43.422 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:43.423 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:43 vm10.local ceph-mon[53712]: pgmap v2718: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:43.448 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:43.449 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:44 vm10.local ceph-mon[53712]: pgmap v2719: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:44.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:44 vm11.local ceph-mon[53973]: pgmap v2719: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:46 vm11.local ceph-mon[53973]: pgmap v2720: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:46 vm10.local ceph-mon[53712]: pgmap v2720: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:48.450 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:48.450 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:48.477 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:48.477 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:49.069 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:48 vm10.local ceph-mon[53712]: pgmap v2721: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:48 vm11.local ceph-mon[53973]: pgmap v2721: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:51.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:50 vm11.local ceph-mon[53973]: pgmap v2722: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:50 vm10.local ceph-mon[53712]: pgmap v2722: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:52 vm11.local ceph-mon[53973]: pgmap v2723: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:52 vm10.local ceph-mon[53712]: pgmap v2723: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:53.479 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:53.480 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:53.506 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:53.506 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:54 vm11.local ceph-mon[53973]: pgmap v2724: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:54 vm10.local ceph-mon[53712]: pgmap v2724: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:56 vm11.local ceph-mon[53973]: pgmap v2725: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:56 vm10.local ceph-mon[53712]: pgmap v2725: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:56:58.508 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:56:58.509 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:56:58.536 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:56:58.536 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:56:59.070 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:56:58 vm10.local ceph-mon[53712]: pgmap v2726: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:56:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:56:58 vm11.local ceph-mon[53973]: pgmap v2726: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:00 vm11.local ceph-mon[53973]: pgmap v2727: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:00 vm10.local ceph-mon[53712]: pgmap v2727: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:02 vm11.local ceph-mon[53973]: pgmap v2728: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:02 vm10.local ceph-mon[53712]: pgmap v2728: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:03.538 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:03.539 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:03.566 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:03.567 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:04 vm11.local ceph-mon[53973]: pgmap v2729: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:04 vm10.local ceph-mon[53712]: pgmap v2729: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:06 vm11.local ceph-mon[53973]: pgmap v2730: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:06 vm10.local ceph-mon[53712]: pgmap v2730: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:08.568 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:08.569 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:08.597 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:08.598 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:08 vm11.local ceph-mon[53973]: pgmap v2731: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:08 vm10.local ceph-mon[53712]: pgmap v2731: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:10 vm10.local ceph-mon[53712]: pgmap v2732: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:10 vm11.local ceph-mon[53973]: pgmap v2732: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:12 vm10.local ceph-mon[53712]: pgmap v2733: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:12 vm11.local ceph-mon[53973]: pgmap v2733: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:13.599 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:13.600 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:13.625 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:13.626 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:14 vm10.local ceph-mon[53712]: pgmap v2734: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:14 vm11.local ceph-mon[53973]: pgmap v2734: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:16 vm10.local ceph-mon[53712]: pgmap v2735: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:16 vm11.local ceph-mon[53973]: pgmap v2735: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:18.627 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:18.628 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:18.654 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:18.654 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:18 vm10.local ceph-mon[53712]: pgmap v2736: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:18 vm11.local ceph-mon[53973]: pgmap v2736: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:20 vm10.local ceph-mon[53712]: pgmap v2737: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:20 vm11.local ceph-mon[53973]: pgmap v2737: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:22 vm10.local ceph-mon[53712]: pgmap v2738: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:22 vm11.local ceph-mon[53973]: pgmap v2738: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:23.656 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:23.656 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:23.683 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:23.684 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:24 vm10.local ceph-mon[53712]: pgmap v2739: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:24 vm11.local ceph-mon[53973]: pgmap v2739: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:26 vm10.local ceph-mon[53712]: pgmap v2740: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:26 vm11.local ceph-mon[53973]: pgmap v2740: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:28.685 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:28.686 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:28.711 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:28.711 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:28 vm10.local ceph-mon[53712]: pgmap v2741: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:57:29.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:57:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:28 vm11.local ceph-mon[53973]: pgmap v2741: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:57:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:57:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:30 vm11.local ceph-mon[53973]: pgmap v2742: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:30 vm10.local ceph-mon[53712]: pgmap v2742: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:32 vm11.local ceph-mon[53973]: pgmap v2743: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:32 vm10.local ceph-mon[53712]: pgmap v2743: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:33.713 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:33.713 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:33.754 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:33.755 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:33.949 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:57:33.949 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:57:33.949 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:57:33.949 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:33 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:57:34.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:57:34.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:57:34.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:57:34.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:33 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:57:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:34 vm11.local ceph-mon[53973]: pgmap v2744: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:57:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:57:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:34 vm10.local ceph-mon[53712]: pgmap v2744: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:57:35.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:57:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:36 vm11.local ceph-mon[53973]: pgmap v2745: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:36 vm10.local ceph-mon[53712]: pgmap v2745: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:38.756 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:38.757 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:38.786 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:38.786 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:38 vm11.local ceph-mon[53973]: pgmap v2746: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:38 vm10.local ceph-mon[53712]: pgmap v2746: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:40 vm11.local ceph-mon[53973]: pgmap v2747: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:40 vm10.local ceph-mon[53712]: pgmap v2747: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:43 vm11.local ceph-mon[53973]: pgmap v2748: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:43 vm10.local ceph-mon[53712]: pgmap v2748: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:43.787 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:43.788 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:43.813 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:43.814 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:45 vm11.local ceph-mon[53973]: pgmap v2749: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:45 vm10.local ceph-mon[53712]: pgmap v2749: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:47 vm10.local ceph-mon[53712]: pgmap v2750: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:47 vm11.local ceph-mon[53973]: pgmap v2750: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:48.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:48 vm10.local ceph-mon[53712]: pgmap v2751: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:48.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:48 vm11.local ceph-mon[53973]: pgmap v2751: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:48.815 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:48.816 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:48.843 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:48.844 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:50.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:50 vm10.local ceph-mon[53712]: pgmap v2752: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:50.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:50 vm11.local ceph-mon[53973]: pgmap v2752: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:52 vm11.local ceph-mon[53973]: pgmap v2753: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:52 vm10.local ceph-mon[53712]: pgmap v2753: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:53.845 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:53.845 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:53.873 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:53.873 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:54 vm11.local ceph-mon[53973]: pgmap v2754: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:54 vm10.local ceph-mon[53712]: pgmap v2754: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:56 vm11.local ceph-mon[53973]: pgmap v2755: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:56 vm10.local ceph-mon[53712]: pgmap v2755: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:57:58.875 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:57:58.875 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:57:58.901 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:57:58.902 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:57:59.074 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:57:58 vm10.local ceph-mon[53712]: pgmap v2756: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:57:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:57:58 vm11.local ceph-mon[53973]: pgmap v2756: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:00 vm10.local ceph-mon[53712]: pgmap v2757: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:00 vm11.local ceph-mon[53973]: pgmap v2757: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:02 vm10.local ceph-mon[53712]: pgmap v2758: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:02 vm11.local ceph-mon[53973]: pgmap v2758: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:03.903 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:03.904 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:03.929 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:03.930 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:04 vm10.local ceph-mon[53712]: pgmap v2759: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:04 vm11.local ceph-mon[53973]: pgmap v2759: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:06 vm10.local ceph-mon[53712]: pgmap v2760: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:06 vm11.local ceph-mon[53973]: pgmap v2760: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:08.931 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:08.931 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:08.960 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:08.960 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:08 vm10.local ceph-mon[53712]: pgmap v2761: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:08 vm11.local ceph-mon[53973]: pgmap v2761: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:10 vm10.local ceph-mon[53712]: pgmap v2762: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:10 vm11.local ceph-mon[53973]: pgmap v2762: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:12 vm10.local ceph-mon[53712]: pgmap v2763: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:12 vm11.local ceph-mon[53973]: pgmap v2763: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:13.962 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:13.962 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:13.989 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:13.989 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:14 vm10.local ceph-mon[53712]: pgmap v2764: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:14 vm11.local ceph-mon[53973]: pgmap v2764: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:16 vm10.local ceph-mon[53712]: pgmap v2765: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:16 vm11.local ceph-mon[53973]: pgmap v2765: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:18.990 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:18.991 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:19.018 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:19.019 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:18 vm10.local ceph-mon[53712]: pgmap v2766: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:18 vm11.local ceph-mon[53973]: pgmap v2766: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:20 vm10.local ceph-mon[53712]: pgmap v2767: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:20 vm11.local ceph-mon[53973]: pgmap v2767: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:22 vm10.local ceph-mon[53712]: pgmap v2768: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:22 vm11.local ceph-mon[53973]: pgmap v2768: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:24.020 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:24.020 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:24.047 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:24.047 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:24 vm11.local ceph-mon[53973]: pgmap v2769: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:24 vm10.local ceph-mon[53712]: pgmap v2769: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:26 vm11.local ceph-mon[53973]: pgmap v2770: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:26 vm10.local ceph-mon[53712]: pgmap v2770: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:29.049 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:29.049 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:29.076 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:29.077 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:28 vm11.local ceph-mon[53973]: pgmap v2771: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:58:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:58:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:28 vm10.local ceph-mon[53712]: pgmap v2771: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:58:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:58:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:30 vm11.local ceph-mon[53973]: pgmap v2772: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:30 vm10.local ceph-mon[53712]: pgmap v2772: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:32 vm11.local ceph-mon[53973]: pgmap v2773: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:32 vm10.local ceph-mon[53712]: pgmap v2773: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:34.078 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:34.079 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:34.108 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:34.108 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: pgmap v2774: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:34 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: pgmap v2774: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:58:35.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:58:35.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:58:35.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:35.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:34 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:58:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:36 vm11.local ceph-mon[53973]: pgmap v2775: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:36 vm10.local ceph-mon[53712]: pgmap v2775: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:39.109 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:39.110 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:39.200 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:39 vm10.local ceph-mon[53712]: pgmap v2776: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:39.203 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:39.204 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:39 vm11.local ceph-mon[53973]: pgmap v2776: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:40.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:40 vm11.local ceph-mon[53973]: pgmap v2777: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:40.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:40 vm10.local ceph-mon[53712]: pgmap v2777: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:42 vm10.local ceph-mon[53712]: pgmap v2778: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:42 vm11.local ceph-mon[53973]: pgmap v2778: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:44.206 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:44.206 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:44.234 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:44.234 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:44 vm10.local ceph-mon[53712]: pgmap v2779: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:44 vm11.local ceph-mon[53973]: pgmap v2779: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:46 vm10.local ceph-mon[53712]: pgmap v2780: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:46 vm11.local ceph-mon[53973]: pgmap v2780: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:48 vm10.local ceph-mon[53712]: pgmap v2781: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:49.236 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:49.237 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:49.265 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:49.265 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:48 vm11.local ceph-mon[53973]: pgmap v2781: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:50 vm10.local ceph-mon[53712]: pgmap v2782: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:50 vm11.local ceph-mon[53973]: pgmap v2782: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:52 vm10.local ceph-mon[53712]: pgmap v2783: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:52 vm11.local ceph-mon[53973]: pgmap v2783: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:54.267 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:54.267 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:54.294 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:54.295 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:54 vm10.local ceph-mon[53712]: pgmap v2784: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:54 vm11.local ceph-mon[53973]: pgmap v2784: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:56 vm10.local ceph-mon[53712]: pgmap v2785: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:56 vm11.local ceph-mon[53973]: pgmap v2785: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:58:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:58:58 vm10.local ceph-mon[53712]: pgmap v2786: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:58:59.296 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:58:59.296 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:58:59.325 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:58:59.325 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:58:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:58:58 vm11.local ceph-mon[53973]: pgmap v2786: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:00 vm10.local ceph-mon[53712]: pgmap v2787: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:00 vm11.local ceph-mon[53973]: pgmap v2787: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:02 vm11.local ceph-mon[53973]: pgmap v2788: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:02 vm10.local ceph-mon[53712]: pgmap v2788: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:04.327 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:04.328 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:04.354 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:04.354 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:04 vm11.local ceph-mon[53973]: pgmap v2789: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:04 vm10.local ceph-mon[53712]: pgmap v2789: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:06 vm11.local ceph-mon[53973]: pgmap v2790: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:07.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:06 vm10.local ceph-mon[53712]: pgmap v2790: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:08 vm11.local ceph-mon[53973]: pgmap v2791: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:09.355 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:09.356 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:09.382 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:09.382 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:08 vm10.local ceph-mon[53712]: pgmap v2791: 97 pgs: 97 active+clean; 453 KiB data, 82 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:11 vm11.local ceph-mon[53973]: pgmap v2792: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:11 vm10.local ceph-mon[53712]: pgmap v2792: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:13 vm10.local ceph-mon[53712]: pgmap v2793: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:13 vm11.local ceph-mon[53973]: pgmap v2793: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:14.384 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:14.384 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:14.411 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:14.412 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:15 vm10.local ceph-mon[53712]: pgmap v2794: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:15 vm11.local ceph-mon[53973]: pgmap v2794: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:17 vm10.local ceph-mon[53712]: pgmap v2795: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:17 vm11.local ceph-mon[53973]: pgmap v2795: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:19.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:19 vm10.local ceph-mon[53712]: pgmap v2796: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:19.413 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:19.413 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:19 vm11.local ceph-mon[53973]: pgmap v2796: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:19.678 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:19.728 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:20.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:20 vm10.local ceph-mon[53712]: pgmap v2797: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:20.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:20 vm11.local ceph-mon[53973]: pgmap v2797: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:22 vm11.local ceph-mon[53973]: pgmap v2798: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:22 vm10.local ceph-mon[53712]: pgmap v2798: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:24.680 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:24.680 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:24.709 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:24.709 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:24 vm11.local ceph-mon[53973]: pgmap v2799: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:24 vm10.local ceph-mon[53712]: pgmap v2799: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:26 vm11.local ceph-mon[53973]: pgmap v2800: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:26 vm10.local ceph-mon[53712]: pgmap v2800: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:28 vm11.local ceph-mon[53973]: pgmap v2801: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:59:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:59:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:28 vm10.local ceph-mon[53712]: pgmap v2801: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T18:59:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T18:59:29.711 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:29.711 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:29.804 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:29.804 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:30 vm11.local ceph-mon[53973]: pgmap v2802: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:30 vm10.local ceph-mon[53712]: pgmap v2802: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:32 vm11.local ceph-mon[53973]: pgmap v2803: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:33.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:32 vm10.local ceph-mon[53712]: pgmap v2803: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:34.805 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:34.806 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:34.838 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:34.838 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:35.038 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:35 vm10.local ceph-mon[53712]: pgmap v2804: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:35.038 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:35 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:59:35.038 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:35 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:59:35.038 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:35 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:59:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:35 vm11.local ceph-mon[53973]: pgmap v2804: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:35 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T18:59:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:35 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T18:59:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:35 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T18:59:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:37 vm11.local ceph-mon[53973]: pgmap v2805: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:37 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:59:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:37 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:59:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:37 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:59:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:37 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:59:37.352 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:37 vm10.local ceph-mon[53712]: pgmap v2805: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:37.352 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:37 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:59:37.352 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:37 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T18:59:37.352 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:37 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:59:37.352 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:37 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T18:59:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:39 vm11.local ceph-mon[53973]: pgmap v2806: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:39 vm10.local ceph-mon[53712]: pgmap v2806: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:39.841 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:39.842 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:39.877 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:39.877 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:41 vm11.local ceph-mon[53973]: pgmap v2807: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:41 vm10.local ceph-mon[53712]: pgmap v2807: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:43 vm10.local ceph-mon[53712]: pgmap v2808: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:43 vm11.local ceph-mon[53973]: pgmap v2808: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:44.879 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:44.879 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:44.907 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:44.907 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:45 vm10.local ceph-mon[53712]: pgmap v2809: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:45 vm11.local ceph-mon[53973]: pgmap v2809: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:47 vm10.local ceph-mon[53712]: pgmap v2810: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:47 vm11.local ceph-mon[53973]: pgmap v2810: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:49.337 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:49 vm10.local ceph-mon[53712]: pgmap v2811: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:49 vm11.local ceph-mon[53973]: pgmap v2811: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:49.908 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:49.909 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:49.946 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:49.946 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:50.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:50 vm10.local ceph-mon[53712]: pgmap v2812: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:50.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:50 vm11.local ceph-mon[53973]: pgmap v2812: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:53 vm11.local ceph-mon[53973]: pgmap v2813: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:53.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:53 vm10.local ceph-mon[53712]: pgmap v2813: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:54.947 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:54.948 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T18:59:54.976 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T18:59:54.976 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T18:59:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:55 vm11.local ceph-mon[53973]: pgmap v2814: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:55 vm10.local ceph-mon[53712]: pgmap v2814: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:57 vm11.local ceph-mon[53973]: pgmap v2815: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:57 vm10.local ceph-mon[53712]: pgmap v2815: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T18:59:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 18:59:59 vm11.local ceph-mon[53973]: pgmap v2816: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 18:59:59 vm10.local ceph-mon[53712]: pgmap v2816: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T18:59:59.978 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T18:59:59.978 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:00.005 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:00.006 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:00.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:00 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T19:00:00.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:00 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T19:00:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:01 vm11.local ceph-mon[53973]: pgmap v2817: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:01 vm10.local ceph-mon[53712]: pgmap v2817: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:03 vm11.local ceph-mon[53973]: pgmap v2818: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:03 vm10.local ceph-mon[53712]: pgmap v2818: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:05.008 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:05.008 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:05.222 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:05.223 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:05 vm10.local ceph-mon[53712]: pgmap v2819: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:05 vm11.local ceph-mon[53973]: pgmap v2819: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:07 vm11.local ceph-mon[53973]: pgmap v2820: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:07 vm10.local ceph-mon[53712]: pgmap v2820: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:09 vm10.local ceph-mon[53712]: pgmap v2821: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:09 vm11.local ceph-mon[53973]: pgmap v2821: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:10.224 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:10.225 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:10.252 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:10.253 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:10.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:10 vm10.local ceph-mon[53712]: pgmap v2822: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:10.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:10 vm11.local ceph-mon[53973]: pgmap v2822: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:13 vm11.local ceph-mon[53973]: pgmap v2823: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:13 vm10.local ceph-mon[53712]: pgmap v2823: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:15.254 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:15.255 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:15.282 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:15.283 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:15 vm11.local ceph-mon[53973]: pgmap v2824: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:15 vm10.local ceph-mon[53712]: pgmap v2824: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:17 vm11.local ceph-mon[53973]: pgmap v2825: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:17 vm10.local ceph-mon[53712]: pgmap v2825: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:19 vm10.local ceph-mon[53712]: pgmap v2826: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:19 vm11.local ceph-mon[53973]: pgmap v2826: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:20.285 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:20.285 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:20.315 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:20.316 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:21 vm11.local ceph-mon[53973]: pgmap v2827: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:21 vm10.local ceph-mon[53712]: pgmap v2827: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:23 vm10.local ceph-mon[53712]: pgmap v2828: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:23 vm11.local ceph-mon[53973]: pgmap v2828: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:24.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:24 vm10.local ceph-mon[53712]: pgmap v2829: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:24.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:24 vm11.local ceph-mon[53973]: pgmap v2829: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:25.318 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:25.318 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:25.345 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:25.345 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:27 vm11.local ceph-mon[53973]: pgmap v2830: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:27 vm10.local ceph-mon[53712]: pgmap v2830: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:29 vm11.local ceph-mon[53973]: pgmap v2831: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:00:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:00:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:29 vm10.local ceph-mon[53712]: pgmap v2831: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:00:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:00:30.347 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:30.347 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:30.372 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:30.373 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:31 vm11.local ceph-mon[53973]: pgmap v2832: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:31 vm10.local ceph-mon[53712]: pgmap v2832: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:33 vm11.local ceph-mon[53973]: pgmap v2833: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:33.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:33 vm10.local ceph-mon[53712]: pgmap v2833: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:35 vm11.local ceph-mon[53973]: pgmap v2834: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:35.374 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:35.374 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:35.400 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:35.401 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:35 vm10.local ceph-mon[53712]: pgmap v2834: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:37.523 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:37 vm11.local ceph-mon[53973]: pgmap v2835: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:37.523 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:37 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:00:37.523 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:37 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:00:37.523 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:37 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:00:37.607 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:37 vm10.local ceph-mon[53712]: pgmap v2835: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:37.607 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:37 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:00:37.608 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:37 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:00:37.608 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:37 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:00:38.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:38 vm11.local ceph-mon[53973]: pgmap v2836: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:38.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:38 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:38.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:38 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:38.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:38 vm10.local ceph-mon[53712]: pgmap v2836: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:38.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:38 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:38.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:38 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:00:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:00:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:00:40.402 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:40.402 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:40.429 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:40.429 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:40 vm11.local ceph-mon[53973]: pgmap v2837: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:40 vm10.local ceph-mon[53712]: pgmap v2837: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:42 vm10.local ceph-mon[53712]: pgmap v2838: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:42 vm11.local ceph-mon[53973]: pgmap v2838: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:44 vm10.local ceph-mon[53712]: pgmap v2839: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:44.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:44 vm11.local ceph-mon[53973]: pgmap v2839: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:45.430 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:45.431 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:45.457 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:45.457 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:46 vm11.local ceph-mon[53973]: pgmap v2840: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:46 vm10.local ceph-mon[53712]: pgmap v2840: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:48.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:48 vm11.local ceph-mon[53973]: pgmap v2841: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:48 vm10.local ceph-mon[53712]: pgmap v2841: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:50.459 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:50.459 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:50.487 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:50.487 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:50.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:50 vm11.local ceph-mon[53973]: pgmap v2842: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:50.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:50 vm10.local ceph-mon[53712]: pgmap v2842: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:52.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:52 vm10.local ceph-mon[53712]: pgmap v2843: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:52 vm11.local ceph-mon[53973]: pgmap v2843: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:54.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:54 vm10.local ceph-mon[53712]: pgmap v2844: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:54 vm11.local ceph-mon[53973]: pgmap v2844: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:55.489 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:00:55.489 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:00:55.516 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:00:55.516 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:00:56.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:56 vm11.local ceph-mon[53973]: pgmap v2845: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:56.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:56 vm10.local ceph-mon[53712]: pgmap v2845: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:00:58.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:00:58 vm11.local ceph-mon[53973]: pgmap v2846: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:00:58.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:00:58 vm10.local ceph-mon[53712]: pgmap v2846: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:00.518 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:00.518 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:00.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:00 vm11.local ceph-mon[53973]: pgmap v2847: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:00.609 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:00.609 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:00.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:00 vm10.local ceph-mon[53712]: pgmap v2847: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:02.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:02 vm10.local ceph-mon[53712]: pgmap v2848: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:02.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:02 vm11.local ceph-mon[53973]: pgmap v2848: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:04.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:04 vm10.local ceph-mon[53712]: pgmap v2849: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:04.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:04 vm11.local ceph-mon[53973]: pgmap v2849: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:05.611 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:05.611 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:05.636 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:05.637 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:06.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:06 vm11.local ceph-mon[53973]: pgmap v2850: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:06.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:06 vm10.local ceph-mon[53712]: pgmap v2850: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:08.422 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:08 vm11.local ceph-mon[53973]: pgmap v2851: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:08.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:08 vm10.local ceph-mon[53712]: pgmap v2851: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:10.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:10 vm11.local ceph-mon[53973]: pgmap v2852: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:10.639 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:10.639 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:10.665 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:10.665 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:10.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:10 vm10.local ceph-mon[53712]: pgmap v2852: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:12.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:12 vm10.local ceph-mon[53712]: pgmap v2853: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:12.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:12 vm11.local ceph-mon[53973]: pgmap v2853: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:14.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:14 vm10.local ceph-mon[53712]: pgmap v2854: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:14.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:14 vm11.local ceph-mon[53973]: pgmap v2854: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:15.667 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:15.667 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:15.694 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:15.694 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:16.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:16 vm11.local ceph-mon[53973]: pgmap v2855: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:16.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:16 vm10.local ceph-mon[53712]: pgmap v2855: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:18.565 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:18 vm11.local ceph-mon[53973]: pgmap v2856: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:18.686 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:18 vm10.local ceph-mon[53712]: pgmap v2856: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:20.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:20 vm10.local ceph-mon[53712]: pgmap v2857: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:20.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:20 vm11.local ceph-mon[53973]: pgmap v2857: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:20.696 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:20.697 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:20.726 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:20.727 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:22.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:22 vm10.local ceph-mon[53712]: pgmap v2858: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:22.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:22 vm11.local ceph-mon[53973]: pgmap v2858: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:24.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:24 vm10.local ceph-mon[53712]: pgmap v2859: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:24.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:24 vm11.local ceph-mon[53973]: pgmap v2859: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:25.728 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:25.728 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:25.755 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:25.756 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:26.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:26 vm11.local ceph-mon[53973]: pgmap v2860: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:26.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:26 vm10.local ceph-mon[53712]: pgmap v2860: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:28.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:28 vm11.local ceph-mon[53973]: pgmap v2861: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:28.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:28 vm10.local ceph-mon[53712]: pgmap v2861: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:01:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:01:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:01:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:01:30.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:30 vm10.local ceph-mon[53712]: pgmap v2862: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:30.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:30 vm11.local ceph-mon[53973]: pgmap v2862: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:30.757 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:30.758 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:30.784 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:30.785 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:32.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:32 vm10.local ceph-mon[53712]: pgmap v2863: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:32.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:32 vm11.local ceph-mon[53973]: pgmap v2863: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:34.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:34 vm10.local ceph-mon[53712]: pgmap v2864: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:34.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:34 vm11.local ceph-mon[53973]: pgmap v2864: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:35.786 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:35.787 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:35.815 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:35.815 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:36.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:36 vm11.local ceph-mon[53973]: pgmap v2865: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:36.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:36 vm10.local ceph-mon[53712]: pgmap v2865: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:38.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:38 vm11.local ceph-mon[53973]: pgmap v2866: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:38.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:38 vm10.local ceph-mon[53712]: pgmap v2866: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:39.223 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:01:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:01:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:01:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:01:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:01:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:01:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:01:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:01:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:01:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:01:40.817 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:40.818 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:40.845 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:40.845 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:41 vm11.local ceph-mon[53973]: pgmap v2867: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:41 vm10.local ceph-mon[53712]: pgmap v2867: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:42 vm10.local ceph-mon[53712]: pgmap v2868: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:42 vm11.local ceph-mon[53973]: pgmap v2868: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:44 vm10.local ceph-mon[53712]: pgmap v2869: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:44.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:44 vm11.local ceph-mon[53973]: pgmap v2869: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:45.846 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:45.847 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:45.872 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:45.873 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:46 vm11.local ceph-mon[53973]: pgmap v2870: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:46 vm10.local ceph-mon[53712]: pgmap v2870: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:48.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:48 vm11.local ceph-mon[53973]: pgmap v2871: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:48 vm10.local ceph-mon[53712]: pgmap v2871: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:50.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:50 vm11.local ceph-mon[53973]: pgmap v2872: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:50.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:50 vm10.local ceph-mon[53712]: pgmap v2872: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:50.874 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:50.875 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:50.903 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:50.903 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:52.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:52 vm10.local ceph-mon[53712]: pgmap v2873: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:52 vm11.local ceph-mon[53973]: pgmap v2873: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:54.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:54 vm10.local ceph-mon[53712]: pgmap v2874: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:54 vm11.local ceph-mon[53973]: pgmap v2874: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:55.905 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:01:55.905 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:01:55.932 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:01:55.932 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:01:56.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:56 vm11.local ceph-mon[53973]: pgmap v2875: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:56.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:56 vm10.local ceph-mon[53712]: pgmap v2875: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:01:58.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:01:58 vm11.local ceph-mon[53973]: pgmap v2876: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:01:58.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:01:58 vm10.local ceph-mon[53712]: pgmap v2876: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:00.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:00 vm11.local ceph-mon[53973]: pgmap v2877: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:00.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:00 vm10.local ceph-mon[53712]: pgmap v2877: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:00.934 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:00.934 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:00.962 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:00.962 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:02.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:02 vm10.local ceph-mon[53712]: pgmap v2878: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:02.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:02 vm11.local ceph-mon[53973]: pgmap v2878: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:04.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:04 vm10.local ceph-mon[53712]: pgmap v2879: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:04.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:04 vm11.local ceph-mon[53973]: pgmap v2879: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:05.964 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:05.964 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:05.991 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:05.991 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:06.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:06 vm11.local ceph-mon[53973]: pgmap v2880: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:06.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:06 vm10.local ceph-mon[53712]: pgmap v2880: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:08.354 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:08 vm11.local ceph-mon[53973]: pgmap v2881: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:08.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:08 vm10.local ceph-mon[53712]: pgmap v2881: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:10.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:10 vm11.local ceph-mon[53973]: pgmap v2882: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:10.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:10 vm10.local ceph-mon[53712]: pgmap v2882: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:10.992 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:10.993 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:11.020 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:11.021 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:12.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:12 vm11.local ceph-mon[53973]: pgmap v2883: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:12.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:12 vm10.local ceph-mon[53712]: pgmap v2883: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:14.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:14 vm11.local ceph-mon[53973]: pgmap v2884: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:14.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:14 vm10.local ceph-mon[53712]: pgmap v2884: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:16.022 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:16.023 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:16.047 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:16.048 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:16.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:16 vm11.local ceph-mon[53973]: pgmap v2885: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:16.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:16 vm10.local ceph-mon[53712]: pgmap v2885: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:18.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:18 vm11.local ceph-mon[53973]: pgmap v2886: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:18.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:18 vm10.local ceph-mon[53712]: pgmap v2886: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:20.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:20 vm10.local ceph-mon[53712]: pgmap v2887: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:20.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:20 vm11.local ceph-mon[53973]: pgmap v2887: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:21.049 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:21.050 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:21.164 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:21.165 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:22.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:22 vm11.local ceph-mon[53973]: pgmap v2888: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:22.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:22 vm10.local ceph-mon[53712]: pgmap v2888: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:24.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:24 vm11.local ceph-mon[53973]: pgmap v2889: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:24.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:24 vm10.local ceph-mon[53712]: pgmap v2889: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:26.167 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:26.167 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:26.196 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:26.197 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:26 vm11.local ceph-mon[53973]: pgmap v2890: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:26 vm10.local ceph-mon[53712]: pgmap v2890: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:28 vm11.local ceph-mon[53973]: pgmap v2891: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:02:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:02:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:28 vm10.local ceph-mon[53712]: pgmap v2891: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:02:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:02:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:30 vm11.local ceph-mon[53973]: pgmap v2892: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:30 vm10.local ceph-mon[53712]: pgmap v2892: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:31.198 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:31.198 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:31.227 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:31.228 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:32 vm11.local ceph-mon[53973]: pgmap v2893: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:32 vm10.local ceph-mon[53712]: pgmap v2893: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:34.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:34 vm10.local ceph-mon[53712]: pgmap v2894: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:34 vm11.local ceph-mon[53973]: pgmap v2894: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:36.229 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:36.229 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:36.256 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:36.257 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:36.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:36 vm10.local ceph-mon[53712]: pgmap v2895: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:37.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:36 vm11.local ceph-mon[53973]: pgmap v2895: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:38 vm10.local ceph-mon[53712]: pgmap v2896: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:38.991 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:38 vm11.local ceph-mon[53973]: pgmap v2896: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:02:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:02:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:02:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:02:39.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:02:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:02:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:02:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:02:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:02:40.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:02:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:40 vm10.local ceph-mon[53712]: pgmap v2897: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:40 vm11.local ceph-mon[53973]: pgmap v2897: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:41.258 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:41.259 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:41.287 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:41.287 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:42.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:42 vm10.local ceph-mon[53712]: pgmap v2898: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:43.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:42 vm11.local ceph-mon[53973]: pgmap v2898: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:44.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:44 vm10.local ceph-mon[53712]: pgmap v2899: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:44 vm11.local ceph-mon[53973]: pgmap v2899: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:46.289 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:46.289 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:46.315 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:46.316 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:46.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:46 vm10.local ceph-mon[53712]: pgmap v2900: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:46 vm11.local ceph-mon[53973]: pgmap v2900: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:48 vm10.local ceph-mon[53712]: pgmap v2901: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:48 vm11.local ceph-mon[53973]: pgmap v2901: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:50.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:50 vm10.local ceph-mon[53712]: pgmap v2902: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:51.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:50 vm11.local ceph-mon[53973]: pgmap v2902: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:51.317 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:51.318 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:51.343 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:51.344 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:52.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:52 vm10.local ceph-mon[53712]: pgmap v2903: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:52 vm11.local ceph-mon[53973]: pgmap v2903: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:54.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:54 vm10.local ceph-mon[53712]: pgmap v2904: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:54 vm11.local ceph-mon[53973]: pgmap v2904: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:56.345 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:02:56.345 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:02:56.371 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:02:56.371 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:02:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:56 vm10.local ceph-mon[53712]: pgmap v2905: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:56 vm11.local ceph-mon[53973]: pgmap v2905: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:02:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:02:58 vm10.local ceph-mon[53712]: pgmap v2906: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:02:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:02:58 vm11.local ceph-mon[53973]: pgmap v2906: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:00.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:00 vm10.local ceph-mon[53712]: pgmap v2907: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:00 vm11.local ceph-mon[53973]: pgmap v2907: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:01.373 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:01.373 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:01.399 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:01.399 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:02.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:02 vm10.local ceph-mon[53712]: pgmap v2908: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:02 vm11.local ceph-mon[53973]: pgmap v2908: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:04.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:04 vm10.local ceph-mon[53712]: pgmap v2909: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:04 vm11.local ceph-mon[53973]: pgmap v2909: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:06.401 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:06.401 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:06.427 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:06.427 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:06.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:06 vm10.local ceph-mon[53712]: pgmap v2910: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:06 vm11.local ceph-mon[53973]: pgmap v2910: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:08.937 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:08 vm10.local ceph-mon[53712]: pgmap v2911: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:08.992 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:08 vm11.local ceph-mon[53973]: pgmap v2911: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:10 vm11.local ceph-mon[53973]: pgmap v2912: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:03:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:10 vm10.local ceph-mon[53712]: pgmap v2912: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:03:11.429 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:11.429 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:11.455 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:11.455 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:12 vm11.local ceph-mon[53973]: pgmap v2913: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:03:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:12 vm10.local ceph-mon[53712]: pgmap v2913: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:03:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:14 vm11.local ceph-mon[53973]: pgmap v2914: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:14 vm10.local ceph-mon[53712]: pgmap v2914: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:16.457 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:16.458 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:16.484 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:16.484 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:17.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:16 vm11.local ceph-mon[53973]: pgmap v2915: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:03:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:16 vm10.local ceph-mon[53712]: pgmap v2915: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:03:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:18 vm11.local ceph-mon[53973]: pgmap v2916: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:19.100 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:18 vm10.local ceph-mon[53712]: pgmap v2916: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:20 vm11.local ceph-mon[53973]: pgmap v2917: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:20 vm10.local ceph-mon[53712]: pgmap v2917: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:21.486 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:21.486 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:21.512 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:21.513 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:22 vm11.local ceph-mon[53973]: pgmap v2918: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:22 vm10.local ceph-mon[53712]: pgmap v2918: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:24 vm11.local ceph-mon[53973]: pgmap v2919: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:24 vm10.local ceph-mon[53712]: pgmap v2919: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:26.514 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:26.514 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:26.539 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:26.539 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:26 vm11.local ceph-mon[53973]: pgmap v2920: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:26 vm10.local ceph-mon[53712]: pgmap v2920: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:28 vm11.local ceph-mon[53973]: pgmap v2921: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:03:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:03:29.101 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:28 vm10.local ceph-mon[53712]: pgmap v2921: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:29.101 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:03:29.101 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:03:31.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:30 vm11.local ceph-mon[53973]: pgmap v2922: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:31.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:30 vm10.local ceph-mon[53712]: pgmap v2922: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:31.541 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:31.541 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:31.567 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:31.568 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:33.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:32 vm11.local ceph-mon[53973]: pgmap v2923: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:33.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:32 vm10.local ceph-mon[53712]: pgmap v2923: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:34 vm11.local ceph-mon[53973]: pgmap v2924: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:34 vm10.local ceph-mon[53712]: pgmap v2924: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:36.569 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:36.570 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:36.596 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:36.596 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:37.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:36 vm11.local ceph-mon[53973]: pgmap v2925: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:37.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:36 vm10.local ceph-mon[53712]: pgmap v2925: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:38 vm11.local ceph-mon[53973]: pgmap v2926: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:39.102 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:38 vm10.local ceph-mon[53712]: pgmap v2926: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:39.803 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:03:39.803 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:03:39.803 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:03:39.803 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:03:39.803 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:39 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:03:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:03:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:03:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:03:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:03:40.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:39 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:03:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:40 vm11.local ceph-mon[53973]: pgmap v2927: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:40 vm10.local ceph-mon[53712]: pgmap v2927: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:41.598 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:41.598 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:41.623 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:41.624 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:43.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:42 vm11.local ceph-mon[53973]: pgmap v2928: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:42 vm10.local ceph-mon[53712]: pgmap v2928: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:44 vm11.local ceph-mon[53973]: pgmap v2929: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:44 vm10.local ceph-mon[53712]: pgmap v2929: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:46.625 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:46.625 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:46.651 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:46.651 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:46 vm11.local ceph-mon[53973]: pgmap v2930: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:46 vm10.local ceph-mon[53712]: pgmap v2930: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:48 vm11.local ceph-mon[53973]: pgmap v2931: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:49.102 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:48 vm10.local ceph-mon[53712]: pgmap v2931: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:51.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:50 vm11.local ceph-mon[53973]: pgmap v2932: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:50 vm10.local ceph-mon[53712]: pgmap v2932: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:51.652 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:51.653 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:51.678 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:51.679 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:52 vm11.local ceph-mon[53973]: pgmap v2933: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:52 vm10.local ceph-mon[53712]: pgmap v2933: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:54 vm10.local ceph-mon[53712]: pgmap v2934: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:54 vm11.local ceph-mon[53973]: pgmap v2934: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:56.680 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:03:56.681 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:03:56.706 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:03:56.707 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:03:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:56 vm10.local ceph-mon[53712]: pgmap v2935: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:56 vm11.local ceph-mon[53973]: pgmap v2935: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:03:59.105 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:03:58 vm10.local ceph-mon[53712]: pgmap v2936: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:03:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:03:58 vm11.local ceph-mon[53973]: pgmap v2936: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:00 vm10.local ceph-mon[53712]: pgmap v2937: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:00 vm11.local ceph-mon[53973]: pgmap v2937: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:01.708 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:01.709 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:01.735 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:01.735 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:02 vm10.local ceph-mon[53712]: pgmap v2938: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:02 vm11.local ceph-mon[53973]: pgmap v2938: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:04 vm10.local ceph-mon[53712]: pgmap v2939: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:04 vm11.local ceph-mon[53973]: pgmap v2939: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:06.737 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:06.738 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:06.764 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:06.765 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:06 vm10.local ceph-mon[53712]: pgmap v2940: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:06 vm11.local ceph-mon[53973]: pgmap v2940: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:08 vm10.local ceph-mon[53712]: pgmap v2941: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:08 vm11.local ceph-mon[53973]: pgmap v2941: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:10 vm10.local ceph-mon[53712]: pgmap v2942: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:10 vm11.local ceph-mon[53973]: pgmap v2942: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:11.766 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:11.767 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:11.793 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:11.794 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:13 vm11.local ceph-mon[53973]: pgmap v2943: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:13 vm10.local ceph-mon[53712]: pgmap v2943: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:15 vm11.local ceph-mon[53973]: pgmap v2944: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:15 vm10.local ceph-mon[53712]: pgmap v2944: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:16.796 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:16.796 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:16.827 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:16.827 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:17 vm11.local ceph-mon[53973]: pgmap v2945: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:17 vm10.local ceph-mon[53712]: pgmap v2945: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:19 vm10.local ceph-mon[53712]: pgmap v2946: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:19 vm11.local ceph-mon[53973]: pgmap v2946: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:21 vm11.local ceph-mon[53973]: pgmap v2947: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:21 vm10.local ceph-mon[53712]: pgmap v2947: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:21.829 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:21.829 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:21.857 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:21.857 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:23 vm11.local ceph-mon[53973]: pgmap v2948: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:23 vm10.local ceph-mon[53712]: pgmap v2948: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:25 vm11.local ceph-mon[53973]: pgmap v2949: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:25.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:25 vm10.local ceph-mon[53712]: pgmap v2949: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:26.859 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:26.859 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:26.885 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:26.885 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:27 vm11.local ceph-mon[53973]: pgmap v2950: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:27 vm10.local ceph-mon[53712]: pgmap v2950: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:29 vm11.local ceph-mon[53973]: pgmap v2951: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:04:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:04:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:29 vm10.local ceph-mon[53712]: pgmap v2951: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:04:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:04:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:31 vm11.local ceph-mon[53973]: pgmap v2952: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:31 vm10.local ceph-mon[53712]: pgmap v2952: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:31.887 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:31.887 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:31.913 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:31.914 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:33 vm11.local ceph-mon[53973]: pgmap v2953: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:33 vm10.local ceph-mon[53712]: pgmap v2953: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:35 vm10.local ceph-mon[53712]: pgmap v2954: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:35 vm11.local ceph-mon[53973]: pgmap v2954: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:36.915 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:36.916 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:36.942 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:36.942 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:37 vm10.local ceph-mon[53712]: pgmap v2955: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:37 vm11.local ceph-mon[53973]: pgmap v2955: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:39 vm10.local ceph-mon[53712]: pgmap v2956: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:39.573 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:39 vm11.local ceph-mon[53973]: pgmap v2956: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:40.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:04:40.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:04:40.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:04:40.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:04:40.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:04:40.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:04:40.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:04:40.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:04:40.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:04:40.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:04:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:41 vm10.local ceph-mon[53712]: pgmap v2957: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:41 vm11.local ceph-mon[53973]: pgmap v2957: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:41.943 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:41.944 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:41.970 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:41.970 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:43 vm10.local ceph-mon[53712]: pgmap v2958: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:43 vm11.local ceph-mon[53973]: pgmap v2958: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:45 vm10.local ceph-mon[53712]: pgmap v2959: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:45 vm11.local ceph-mon[53973]: pgmap v2959: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:46.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:46 vm10.local ceph-mon[53712]: pgmap v2960: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:46 vm11.local ceph-mon[53973]: pgmap v2960: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:46.972 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:46.972 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:46.997 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:46.997 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:48.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:48 vm11.local ceph-mon[53973]: pgmap v2961: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:48 vm10.local ceph-mon[53712]: pgmap v2961: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:50.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:50 vm11.local ceph-mon[53973]: pgmap v2962: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:50.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:50 vm10.local ceph-mon[53712]: pgmap v2962: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:51.998 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:51.999 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:52.025 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:52.026 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:52.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:52 vm11.local ceph-mon[53973]: pgmap v2963: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:52.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:52 vm10.local ceph-mon[53712]: pgmap v2963: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:54.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:54 vm10.local ceph-mon[53712]: pgmap v2964: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:54 vm11.local ceph-mon[53973]: pgmap v2964: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:56.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:56 vm11.local ceph-mon[53973]: pgmap v2965: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:56.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:56 vm10.local ceph-mon[53712]: pgmap v2965: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:04:57.027 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:04:57.028 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:04:57.055 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:04:57.055 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:04:58.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:04:58 vm11.local ceph-mon[53973]: pgmap v2966: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:04:58.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:04:58 vm10.local ceph-mon[53712]: pgmap v2966: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:00.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:00 vm11.local ceph-mon[53973]: pgmap v2967: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:00.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:00 vm10.local ceph-mon[53712]: pgmap v2967: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:02.057 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:02.057 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:02.085 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:02.086 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:02.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:02 vm11.local ceph-mon[53973]: pgmap v2968: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:02.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:02 vm10.local ceph-mon[53712]: pgmap v2968: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:04.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:04 vm11.local ceph-mon[53973]: pgmap v2969: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:04.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:04 vm10.local ceph-mon[53712]: pgmap v2969: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:06.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:06 vm11.local ceph-mon[53973]: pgmap v2970: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:06.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:06 vm10.local ceph-mon[53712]: pgmap v2970: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:07.087 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:07.088 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:07.113 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:07.114 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:08.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:08 vm11.local ceph-mon[53973]: pgmap v2971: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:08.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:08 vm10.local ceph-mon[53712]: pgmap v2971: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:10.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:10 vm11.local ceph-mon[53973]: pgmap v2972: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:10.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:10 vm10.local ceph-mon[53712]: pgmap v2972: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:12.115 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:12.115 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:12.143 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:12.144 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:12.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:12 vm11.local ceph-mon[53973]: pgmap v2973: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:12.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:12 vm10.local ceph-mon[53712]: pgmap v2973: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:14.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:14 vm11.local ceph-mon[53973]: pgmap v2974: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:14.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:14 vm10.local ceph-mon[53712]: pgmap v2974: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:16.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:16 vm11.local ceph-mon[53973]: pgmap v2975: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:16.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:16 vm10.local ceph-mon[53712]: pgmap v2975: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:17.145 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:17.145 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:17.174 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:17.174 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:18.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:18 vm11.local ceph-mon[53973]: pgmap v2976: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:18 vm10.local ceph-mon[53712]: pgmap v2976: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:20.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:20 vm11.local ceph-mon[53973]: pgmap v2977: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:20.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:20 vm10.local ceph-mon[53712]: pgmap v2977: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:22.175 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:22.176 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:22.202 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:22.203 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:22.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:22 vm11.local ceph-mon[53973]: pgmap v2978: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:22.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:22 vm10.local ceph-mon[53712]: pgmap v2978: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:24.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:24 vm11.local ceph-mon[53973]: pgmap v2979: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:24.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:24 vm10.local ceph-mon[53712]: pgmap v2979: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:26 vm11.local ceph-mon[53973]: pgmap v2980: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:26 vm10.local ceph-mon[53712]: pgmap v2980: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:27.205 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:27.205 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:27.231 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:27.232 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:28 vm11.local ceph-mon[53973]: pgmap v2981: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:05:28.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:05:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:28 vm10.local ceph-mon[53712]: pgmap v2981: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:05:28.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:05:30.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:30 vm11.local ceph-mon[53973]: pgmap v2982: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:30.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:30 vm10.local ceph-mon[53712]: pgmap v2982: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:32.233 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:32.234 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:32.261 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:32.261 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:32 vm11.local ceph-mon[53973]: pgmap v2983: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:32 vm10.local ceph-mon[53712]: pgmap v2983: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:34.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:34 vm11.local ceph-mon[53973]: pgmap v2984: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:34.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:34 vm10.local ceph-mon[53712]: pgmap v2984: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:36.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:36 vm11.local ceph-mon[53973]: pgmap v2985: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:36.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:36 vm10.local ceph-mon[53712]: pgmap v2985: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:37.263 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:37.264 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:37.290 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:37.290 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:38 vm11.local ceph-mon[53973]: pgmap v2986: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:38 vm10.local ceph-mon[53712]: pgmap v2986: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:40 vm11.local ceph-mon[53973]: pgmap v2987: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:05:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:05:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:05:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:05:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:05:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:40 vm10.local ceph-mon[53712]: pgmap v2987: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:05:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:05:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:05:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:05:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:05:42.292 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:42.292 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:42.319 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:42.320 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:42.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:42 vm11.local ceph-mon[53973]: pgmap v2988: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:42.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:42 vm10.local ceph-mon[53712]: pgmap v2988: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:44.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:44 vm11.local ceph-mon[53973]: pgmap v2989: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:44.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:44 vm10.local ceph-mon[53712]: pgmap v2989: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:46.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:46 vm11.local ceph-mon[53973]: pgmap v2990: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:46.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:46 vm10.local ceph-mon[53712]: pgmap v2990: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:47.321 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:47.322 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:47.351 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:47.351 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:48 vm11.local ceph-mon[53973]: pgmap v2991: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:48 vm10.local ceph-mon[53712]: pgmap v2991: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:50.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:50 vm11.local ceph-mon[53973]: pgmap v2992: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:50.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:50 vm10.local ceph-mon[53712]: pgmap v2992: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:52.353 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:52.353 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:52.383 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:52.383 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:52.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:52 vm11.local ceph-mon[53973]: pgmap v2993: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:52.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:52 vm10.local ceph-mon[53712]: pgmap v2993: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:54.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:54 vm11.local ceph-mon[53973]: pgmap v2994: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:54.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:54 vm10.local ceph-mon[53712]: pgmap v2994: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:56 vm10.local ceph-mon[53712]: pgmap v2995: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:56 vm11.local ceph-mon[53973]: pgmap v2995: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:05:57.385 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:05:57.386 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:05:57.417 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:05:57.417 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:05:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:05:58 vm10.local ceph-mon[53712]: pgmap v2996: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:05:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:05:58 vm11.local ceph-mon[53973]: pgmap v2996: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:00.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:00 vm10.local ceph-mon[53712]: pgmap v2997: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:00 vm11.local ceph-mon[53973]: pgmap v2997: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:02.419 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:02.419 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:02.446 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:02.446 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:02.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:02 vm10.local ceph-mon[53712]: pgmap v2998: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:02 vm11.local ceph-mon[53973]: pgmap v2998: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:04.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:04 vm10.local ceph-mon[53712]: pgmap v2999: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:04 vm11.local ceph-mon[53973]: pgmap v2999: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:06.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:06 vm10.local ceph-mon[53712]: pgmap v3000: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:06 vm11.local ceph-mon[53973]: pgmap v3000: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:07.448 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:07.448 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:07.475 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:07.475 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:08 vm10.local ceph-mon[53712]: pgmap v3001: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:08.992 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:08 vm11.local ceph-mon[53973]: pgmap v3001: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:10 vm10.local ceph-mon[53712]: pgmap v3002: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:10 vm11.local ceph-mon[53973]: pgmap v3002: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:12.477 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:12.478 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:12.505 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:12.505 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:12 vm10.local ceph-mon[53712]: pgmap v3003: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:12 vm11.local ceph-mon[53973]: pgmap v3003: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:14.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:14 vm10.local ceph-mon[53712]: pgmap v3004: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:14 vm11.local ceph-mon[53973]: pgmap v3004: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:16.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:16 vm10.local ceph-mon[53712]: pgmap v3005: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:17.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:16 vm11.local ceph-mon[53973]: pgmap v3005: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:17.507 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:17.507 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:17.536 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:17.536 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:18 vm10.local ceph-mon[53712]: pgmap v3006: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:18 vm11.local ceph-mon[53973]: pgmap v3006: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:20.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:20 vm10.local ceph-mon[53712]: pgmap v3007: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:20 vm11.local ceph-mon[53973]: pgmap v3007: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:22.538 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:22.538 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:22.566 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:22.566 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:22.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:22 vm10.local ceph-mon[53712]: pgmap v3008: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:22 vm11.local ceph-mon[53973]: pgmap v3008: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:24.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:24 vm10.local ceph-mon[53712]: pgmap v3009: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:24 vm11.local ceph-mon[53973]: pgmap v3009: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:26 vm11.local ceph-mon[53973]: pgmap v3010: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:26 vm10.local ceph-mon[53712]: pgmap v3010: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:27.568 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:27.568 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:27.595 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:27.595 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:28 vm11.local ceph-mon[53973]: pgmap v3011: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:06:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:06:29.115 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:28 vm10.local ceph-mon[53712]: pgmap v3011: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:29.115 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:06:29.115 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:06:31.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:30 vm11.local ceph-mon[53973]: pgmap v3012: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:31.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:30 vm10.local ceph-mon[53712]: pgmap v3012: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:32.597 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:32.597 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:32.624 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:32.625 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:33.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:32 vm11.local ceph-mon[53973]: pgmap v3013: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:33.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:32 vm10.local ceph-mon[53712]: pgmap v3013: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:34 vm11.local ceph-mon[53973]: pgmap v3014: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:34 vm10.local ceph-mon[53712]: pgmap v3014: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:37.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:36 vm11.local ceph-mon[53973]: pgmap v3015: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:37.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:36 vm10.local ceph-mon[53712]: pgmap v3015: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:37.626 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:37.627 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:37.656 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:37.657 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:38 vm11.local ceph-mon[53973]: pgmap v3016: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:39.116 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:38 vm10.local ceph-mon[53712]: pgmap v3016: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:40 vm11.local ceph-mon[53973]: pgmap v3017: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:06:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:06:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:06:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:06:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:06:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:40 vm10.local ceph-mon[53712]: pgmap v3017: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:06:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:06:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:06:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:06:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:06:42.658 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:42.659 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:42.687 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:42.687 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:43.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:42 vm11.local ceph-mon[53973]: pgmap v3018: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:42 vm10.local ceph-mon[53712]: pgmap v3018: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:44 vm11.local ceph-mon[53973]: pgmap v3019: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:44 vm10.local ceph-mon[53712]: pgmap v3019: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:46 vm11.local ceph-mon[53973]: pgmap v3020: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:46 vm10.local ceph-mon[53712]: pgmap v3020: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:47.688 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:47.689 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:47.715 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:47.716 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:48 vm11.local ceph-mon[53973]: pgmap v3021: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:49.116 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:48 vm10.local ceph-mon[53712]: pgmap v3021: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:50 vm10.local ceph-mon[53712]: pgmap v3022: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:50 vm11.local ceph-mon[53973]: pgmap v3022: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:52.718 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:52.718 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:52.746 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:52.747 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:52 vm10.local ceph-mon[53712]: pgmap v3023: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:52 vm11.local ceph-mon[53973]: pgmap v3023: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:54 vm10.local ceph-mon[53712]: pgmap v3024: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:54 vm11.local ceph-mon[53973]: pgmap v3024: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:56 vm10.local ceph-mon[53712]: pgmap v3025: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:56 vm11.local ceph-mon[53973]: pgmap v3025: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:06:57.748 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:06:57.749 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:06:57.777 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:06:57.777 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:06:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:06:58 vm10.local ceph-mon[53712]: pgmap v3026: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:06:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:06:58 vm11.local ceph-mon[53973]: pgmap v3026: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:00 vm10.local ceph-mon[53712]: pgmap v3027: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:00 vm11.local ceph-mon[53973]: pgmap v3027: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:02.779 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:02.779 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:02.823 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:02.823 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:02 vm10.local ceph-mon[53712]: pgmap v3028: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:02 vm11.local ceph-mon[53973]: pgmap v3028: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:04 vm10.local ceph-mon[53712]: pgmap v3029: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:04 vm11.local ceph-mon[53973]: pgmap v3029: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:06 vm10.local ceph-mon[53712]: pgmap v3030: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:06 vm11.local ceph-mon[53973]: pgmap v3030: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:07.825 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:07.825 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:07.853 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:07.853 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:08 vm10.local ceph-mon[53712]: pgmap v3031: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:08 vm11.local ceph-mon[53973]: pgmap v3031: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:10 vm10.local ceph-mon[53712]: pgmap v3032: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:10 vm11.local ceph-mon[53973]: pgmap v3032: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:12.855 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:12.856 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:12.883 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:12.883 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:12 vm10.local ceph-mon[53712]: pgmap v3033: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:12 vm11.local ceph-mon[53973]: pgmap v3033: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:14 vm10.local ceph-mon[53712]: pgmap v3034: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:14 vm11.local ceph-mon[53973]: pgmap v3034: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:16 vm11.local ceph-mon[53973]: pgmap v3035: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:16 vm10.local ceph-mon[53712]: pgmap v3035: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:17.885 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:17.885 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:17.921 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:17.921 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:19.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:18 vm10.local ceph-mon[53712]: pgmap v3036: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:18 vm11.local ceph-mon[53973]: pgmap v3036: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:21 vm11.local ceph-mon[53973]: pgmap v3037: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:21.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:21 vm10.local ceph-mon[53712]: pgmap v3037: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:22.923 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:22.923 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:22.951 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:22.952 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:23 vm11.local ceph-mon[53973]: pgmap v3038: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:23 vm10.local ceph-mon[53712]: pgmap v3038: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:25 vm11.local ceph-mon[53973]: pgmap v3039: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:25 vm10.local ceph-mon[53712]: pgmap v3039: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:27 vm10.local ceph-mon[53712]: pgmap v3040: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:27 vm11.local ceph-mon[53973]: pgmap v3040: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:27.954 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:27.954 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:28.179 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:28.180 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:28.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:28 vm11.local ceph-mon[53973]: pgmap v3041: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:28.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:28 vm10.local ceph-mon[53712]: pgmap v3041: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:07:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:07:29.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:07:29.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:07:30.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:30 vm11.local ceph-mon[53973]: pgmap v3042: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:30.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:30 vm10.local ceph-mon[53712]: pgmap v3042: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:32.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:32 vm11.local ceph-mon[53973]: pgmap v3043: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:32.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:32 vm10.local ceph-mon[53712]: pgmap v3043: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:33.181 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:33.182 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:33.209 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:33.209 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:34.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:34 vm11.local ceph-mon[53973]: pgmap v3044: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:34.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:34 vm10.local ceph-mon[53712]: pgmap v3044: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:36.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:36 vm11.local ceph-mon[53973]: pgmap v3045: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:36.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:36 vm10.local ceph-mon[53712]: pgmap v3045: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:38.211 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:38.211 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:38.238 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:38.239 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:38.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:38 vm11.local ceph-mon[53973]: pgmap v3046: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:38 vm10.local ceph-mon[53712]: pgmap v3046: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:40.611 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:40 vm11.local ceph-mon[53973]: pgmap v3047: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:40.611 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:07:40.611 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:07:40.611 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:40 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:07:40.642 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:40 vm10.local ceph-mon[53712]: pgmap v3047: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:40.642 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:07:40.642 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:07:40.642 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:40 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:07:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:41 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:07:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:41 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:07:41.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:41 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:07:41.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:41 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:07:41.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:41 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:07:41.935 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:41 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:07:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:42 vm10.local ceph-mon[53712]: pgmap v3048: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:43.240 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:43.240 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:43.267 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:43.268 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:42 vm11.local ceph-mon[53973]: pgmap v3048: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:44 vm10.local ceph-mon[53712]: pgmap v3049: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:44 vm11.local ceph-mon[53973]: pgmap v3049: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:46 vm10.local ceph-mon[53712]: pgmap v3050: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:46 vm11.local ceph-mon[53973]: pgmap v3050: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:48.269 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:48.270 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:48.296 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:48.296 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:49.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:48 vm10.local ceph-mon[53712]: pgmap v3051: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:48 vm11.local ceph-mon[53973]: pgmap v3051: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:50 vm10.local ceph-mon[53712]: pgmap v3052: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:50 vm11.local ceph-mon[53973]: pgmap v3052: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:52 vm10.local ceph-mon[53712]: pgmap v3053: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:53.298 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:53.299 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:53.326 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:53.327 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:52 vm11.local ceph-mon[53973]: pgmap v3053: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:54 vm10.local ceph-mon[53712]: pgmap v3054: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:54 vm11.local ceph-mon[53973]: pgmap v3054: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:56 vm10.local ceph-mon[53712]: pgmap v3055: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:56 vm11.local ceph-mon[53973]: pgmap v3055: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:07:58.329 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:07:58.329 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:07:58.356 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:07:58.357 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:07:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:07:58 vm11.local ceph-mon[53973]: pgmap v3056: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:07:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:07:58 vm10.local ceph-mon[53712]: pgmap v3056: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:00 vm11.local ceph-mon[53973]: pgmap v3057: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:00 vm10.local ceph-mon[53712]: pgmap v3057: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:02 vm11.local ceph-mon[53973]: pgmap v3058: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:03.358 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:03.359 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:03.392 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:03.393 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:02 vm10.local ceph-mon[53712]: pgmap v3058: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:04 vm11.local ceph-mon[53973]: pgmap v3059: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:04 vm10.local ceph-mon[53712]: pgmap v3059: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:06 vm11.local ceph-mon[53973]: pgmap v3060: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:06 vm10.local ceph-mon[53712]: pgmap v3060: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:08.394 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:08.395 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:08.420 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:08.421 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:08 vm11.local ceph-mon[53973]: pgmap v3061: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:08 vm10.local ceph-mon[53712]: pgmap v3061: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:10 vm11.local ceph-mon[53973]: pgmap v3062: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:10 vm10.local ceph-mon[53712]: pgmap v3062: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:12 vm11.local ceph-mon[53973]: pgmap v3063: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:13.422 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:13.423 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:12 vm10.local ceph-mon[53712]: pgmap v3063: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:13.449 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:13.449 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:14 vm11.local ceph-mon[53973]: pgmap v3064: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:14 vm10.local ceph-mon[53712]: pgmap v3064: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:16 vm11.local ceph-mon[53973]: pgmap v3065: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:16 vm10.local ceph-mon[53712]: pgmap v3065: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:18.450 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:18.451 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:18.477 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:18.478 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:19.238 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:18 vm10.local ceph-mon[53712]: pgmap v3066: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:19 vm11.local ceph-mon[53973]: pgmap v3066: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:21 vm11.local ceph-mon[53973]: pgmap v3067: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:21 vm10.local ceph-mon[53712]: pgmap v3067: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:23 vm11.local ceph-mon[53973]: pgmap v3068: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:23 vm10.local ceph-mon[53712]: pgmap v3068: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:23.480 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:23.480 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:23.506 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:23.507 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:25 vm11.local ceph-mon[53973]: pgmap v3069: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:25 vm10.local ceph-mon[53712]: pgmap v3069: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:27 vm11.local ceph-mon[53973]: pgmap v3070: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:27 vm10.local ceph-mon[53712]: pgmap v3070: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:28.508 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:28.509 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:28.535 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:28.535 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:29 vm11.local ceph-mon[53973]: pgmap v3071: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:08:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:08:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:29 vm10.local ceph-mon[53712]: pgmap v3071: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:08:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:08:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:31 vm11.local ceph-mon[53973]: pgmap v3072: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:31 vm10.local ceph-mon[53712]: pgmap v3072: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:33 vm11.local ceph-mon[53973]: pgmap v3073: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:33 vm10.local ceph-mon[53712]: pgmap v3073: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:33.537 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:33.537 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:33.564 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:33.564 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:35 vm11.local ceph-mon[53973]: pgmap v3074: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:35 vm10.local ceph-mon[53712]: pgmap v3074: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:37 vm10.local ceph-mon[53712]: pgmap v3075: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:37 vm11.local ceph-mon[53973]: pgmap v3075: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:38.566 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:38.566 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:38.593 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:38.593 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:39 vm10.local ceph-mon[53712]: pgmap v3076: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:39 vm11.local ceph-mon[53973]: pgmap v3076: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:41.172 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:41 vm11.local ceph-mon[53973]: pgmap v3077: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:41.172 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:41 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:08:41.172 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:41 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:08:41.172 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:41 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:08:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:41 vm10.local ceph-mon[53712]: pgmap v3077: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:41.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:41 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:08:41.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:41 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:08:41.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:41 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:08:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:08:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:08:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:08:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:08:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:08:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:08:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:43 vm10.local ceph-mon[53712]: pgmap v3078: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:43 vm11.local ceph-mon[53973]: pgmap v3078: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:43.594 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:43.595 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:43.621 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:43.622 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:45 vm10.local ceph-mon[53712]: pgmap v3079: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:45 vm11.local ceph-mon[53973]: pgmap v3079: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:47 vm10.local ceph-mon[53712]: pgmap v3080: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:47 vm11.local ceph-mon[53973]: pgmap v3080: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:48.624 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:48.624 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:48.649 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:48.650 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:49 vm10.local ceph-mon[53712]: pgmap v3081: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:49 vm11.local ceph-mon[53973]: pgmap v3081: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:51 vm10.local ceph-mon[53712]: pgmap v3082: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:51 vm11.local ceph-mon[53973]: pgmap v3082: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:52.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:52 vm10.local ceph-mon[53712]: pgmap v3083: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:52.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:52 vm11.local ceph-mon[53973]: pgmap v3083: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:53.651 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:53.652 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:53.695 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:53.696 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:54.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:54 vm10.local ceph-mon[53712]: pgmap v3084: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:54.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:54 vm11.local ceph-mon[53973]: pgmap v3084: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:56 vm11.local ceph-mon[53973]: pgmap v3085: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:56 vm10.local ceph-mon[53712]: pgmap v3085: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:08:58.698 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:08:58.699 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:08:58.727 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:08:58.728 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:08:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:08:58 vm10.local ceph-mon[53712]: pgmap v3086: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:08:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:08:58 vm11.local ceph-mon[53973]: pgmap v3086: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:00.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:00 vm10.local ceph-mon[53712]: pgmap v3087: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:00 vm11.local ceph-mon[53973]: pgmap v3087: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:02.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:02 vm10.local ceph-mon[53712]: pgmap v3088: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:02 vm11.local ceph-mon[53973]: pgmap v3088: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:03.729 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:03.730 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:03.787 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:03.787 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:04 vm11.local ceph-mon[53973]: pgmap v3089: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:04 vm10.local ceph-mon[53712]: pgmap v3089: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:06 vm11.local ceph-mon[53973]: pgmap v3090: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:06 vm10.local ceph-mon[53712]: pgmap v3090: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:08.789 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:08.789 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:08.822 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:08.823 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:08 vm11.local ceph-mon[53973]: pgmap v3091: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:09.129 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:08 vm10.local ceph-mon[53712]: pgmap v3091: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:10 vm11.local ceph-mon[53973]: pgmap v3092: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:10 vm10.local ceph-mon[53712]: pgmap v3092: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:12 vm10.local ceph-mon[53712]: pgmap v3093: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:12 vm11.local ceph-mon[53973]: pgmap v3093: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:13.824 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:13.825 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:13.852 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:13.852 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:14 vm10.local ceph-mon[53712]: pgmap v3094: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:14 vm11.local ceph-mon[53973]: pgmap v3094: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:16 vm10.local ceph-mon[53712]: pgmap v3095: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:16 vm11.local ceph-mon[53973]: pgmap v3095: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:18.854 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:18.854 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:18.892 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:18.892 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:18 vm10.local ceph-mon[53712]: pgmap v3096: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:18 vm11.local ceph-mon[53973]: pgmap v3096: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:20 vm10.local ceph-mon[53712]: pgmap v3097: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:20 vm11.local ceph-mon[53973]: pgmap v3097: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:22 vm11.local ceph-mon[53973]: pgmap v3098: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:22 vm10.local ceph-mon[53712]: pgmap v3098: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:23.894 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:23.894 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:23.922 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:23.922 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:24 vm11.local ceph-mon[53973]: pgmap v3099: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:24 vm10.local ceph-mon[53712]: pgmap v3099: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:26 vm11.local ceph-mon[53973]: pgmap v3100: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:26 vm10.local ceph-mon[53712]: pgmap v3100: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:28.924 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:28.925 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:28.952 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:28.952 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:28 vm11.local ceph-mon[53973]: pgmap v3101: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:09:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:09:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:28 vm10.local ceph-mon[53712]: pgmap v3101: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:09:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:09:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:30 vm11.local ceph-mon[53973]: pgmap v3102: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:30 vm10.local ceph-mon[53712]: pgmap v3102: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:32 vm11.local ceph-mon[53973]: pgmap v3103: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:32 vm10.local ceph-mon[53712]: pgmap v3103: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:33.954 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:33.954 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:33.981 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:33.981 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:35 vm11.local ceph-mon[53973]: pgmap v3104: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:35 vm10.local ceph-mon[53712]: pgmap v3104: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:37 vm11.local ceph-mon[53973]: pgmap v3105: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:37 vm10.local ceph-mon[53712]: pgmap v3105: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:38.983 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:38.983 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:39.028 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:39.028 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:39 vm11.local ceph-mon[53973]: pgmap v3106: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:39 vm10.local ceph-mon[53712]: pgmap v3106: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:41 vm11.local ceph-mon[53973]: pgmap v3107: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:41.429 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:41 vm10.local ceph-mon[53712]: pgmap v3107: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:09:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:09:42.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:09:42.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:09:42.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:09:42.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:09:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:09:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:09:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:09:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:09:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:09:42.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:09:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:43 vm11.local ceph-mon[53973]: pgmap v3108: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:43 vm10.local ceph-mon[53712]: pgmap v3108: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:44.030 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:44.030 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:44.276 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:44.277 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:44.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:44 vm11.local ceph-mon[53973]: pgmap v3109: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:44.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:44 vm10.local ceph-mon[53712]: pgmap v3109: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:46 vm11.local ceph-mon[53973]: pgmap v3110: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:46 vm10.local ceph-mon[53712]: pgmap v3110: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:48.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:48 vm11.local ceph-mon[53973]: pgmap v3111: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:48.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:48 vm10.local ceph-mon[53712]: pgmap v3111: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:49.279 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:49.279 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:49.305 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:49.306 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:50.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:50 vm11.local ceph-mon[53973]: pgmap v3112: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:50.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:50 vm10.local ceph-mon[53712]: pgmap v3112: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:52.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:52 vm11.local ceph-mon[53973]: pgmap v3113: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:52.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:52 vm10.local ceph-mon[53712]: pgmap v3113: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:54.307 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:54.308 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:54.335 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:54.336 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:09:54.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:54 vm11.local ceph-mon[53973]: pgmap v3114: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:54.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:54 vm10.local ceph-mon[53712]: pgmap v3114: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:56.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:56 vm11.local ceph-mon[53973]: pgmap v3115: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:56.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:56 vm10.local ceph-mon[53712]: pgmap v3115: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:09:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:09:58 vm10.local ceph-mon[53712]: pgmap v3116: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:09:58 vm11.local ceph-mon[53973]: pgmap v3116: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:09:59.337 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:09:59.338 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:09:59.364 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:09:59.364 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:00.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:00 vm10.local ceph-mon[53712]: pgmap v3117: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:00.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:00 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T19:10:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:00 vm11.local ceph-mon[53973]: pgmap v3117: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:00 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T19:10:02.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:02 vm10.local ceph-mon[53712]: pgmap v3118: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:02 vm11.local ceph-mon[53973]: pgmap v3118: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:04.365 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:04.366 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:04.408 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:04.409 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:04.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:04 vm10.local ceph-mon[53712]: pgmap v3119: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:04 vm11.local ceph-mon[53973]: pgmap v3119: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:06.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:06 vm10.local ceph-mon[53712]: pgmap v3120: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:06 vm11.local ceph-mon[53973]: pgmap v3120: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:08 vm10.local ceph-mon[53712]: pgmap v3121: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:08.991 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:08 vm11.local ceph-mon[53973]: pgmap v3121: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:09.410 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:09.411 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:09.439 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:09.439 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:10 vm10.local ceph-mon[53712]: pgmap v3122: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:10 vm11.local ceph-mon[53973]: pgmap v3122: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:12 vm10.local ceph-mon[53712]: pgmap v3123: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:12 vm11.local ceph-mon[53973]: pgmap v3123: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:14.440 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:14.441 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:14.466 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:14.466 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:14.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:14 vm10.local ceph-mon[53712]: pgmap v3124: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:14 vm11.local ceph-mon[53973]: pgmap v3124: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:16.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:16 vm10.local ceph-mon[53712]: pgmap v3125: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:17.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:16 vm11.local ceph-mon[53973]: pgmap v3125: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:18 vm11.local ceph-mon[53973]: pgmap v3126: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:19.135 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:18 vm10.local ceph-mon[53712]: pgmap v3126: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:19.468 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:19.468 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:19.498 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:19.499 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:20 vm11.local ceph-mon[53973]: pgmap v3127: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:20 vm10.local ceph-mon[53712]: pgmap v3127: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:22 vm11.local ceph-mon[53973]: pgmap v3128: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:22 vm10.local ceph-mon[53712]: pgmap v3128: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:24.500 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:24.501 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:24.530 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:24.530 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:24 vm11.local ceph-mon[53973]: pgmap v3129: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:24 vm10.local ceph-mon[53712]: pgmap v3129: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:26 vm11.local ceph-mon[53973]: pgmap v3130: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:26 vm10.local ceph-mon[53712]: pgmap v3130: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:28 vm11.local ceph-mon[53973]: pgmap v3131: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:10:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:10:29.136 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:28 vm10.local ceph-mon[53712]: pgmap v3131: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:29.136 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:10:29.136 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:10:29.532 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:29.532 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:29.559 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:29.560 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:31.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:30 vm11.local ceph-mon[53973]: pgmap v3132: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:31.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:30 vm10.local ceph-mon[53712]: pgmap v3132: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:33 vm11.local ceph-mon[53973]: pgmap v3133: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:33.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:32 vm10.local ceph-mon[53712]: pgmap v3133: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:34.561 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:34.562 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:34.589 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:34.589 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:34 vm11.local ceph-mon[53973]: pgmap v3134: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:34 vm10.local ceph-mon[53712]: pgmap v3134: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:36 vm11.local ceph-mon[53973]: pgmap v3135: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:36 vm10.local ceph-mon[53712]: pgmap v3135: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:39 vm11.local ceph-mon[53973]: pgmap v3136: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:39 vm10.local ceph-mon[53712]: pgmap v3136: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:39.591 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:39.591 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:39.618 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:39.618 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:41 vm11.local ceph-mon[53973]: pgmap v3137: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:41 vm10.local ceph-mon[53712]: pgmap v3137: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:42.761 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:42 vm10.local ceph-mon[53712]: pgmap v3138: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:42.761 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:10:42.761 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:10:42.761 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:42 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:10:42.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:42 vm11.local ceph-mon[53973]: pgmap v3138: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:42.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:10:42.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:10:42.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:42 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:10:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:10:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:10:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:10:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:10:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:10:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:10:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:10:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:10:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:10:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:10:44.620 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:44.621 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:44.649 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:44.650 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:45 vm10.local ceph-mon[53712]: pgmap v3139: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:45 vm11.local ceph-mon[53973]: pgmap v3139: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:47 vm10.local ceph-mon[53712]: pgmap v3140: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:47 vm11.local ceph-mon[53973]: pgmap v3140: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:49 vm10.local ceph-mon[53712]: pgmap v3141: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:49 vm11.local ceph-mon[53973]: pgmap v3141: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:49.651 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:49.652 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:49.681 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:49.681 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:51 vm10.local ceph-mon[53712]: pgmap v3142: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:51 vm11.local ceph-mon[53973]: pgmap v3142: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:53 vm11.local ceph-mon[53973]: pgmap v3143: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:53 vm10.local ceph-mon[53712]: pgmap v3143: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:54.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:54 vm10.local ceph-mon[53712]: pgmap v3144: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:54.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:54 vm11.local ceph-mon[53973]: pgmap v3144: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:54.683 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:54.683 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:54.738 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:54.739 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:10:56.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:56 vm11.local ceph-mon[53973]: pgmap v3145: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:56.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:56 vm10.local ceph-mon[53712]: pgmap v3145: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:10:58.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:10:58 vm10.local ceph-mon[53712]: pgmap v3146: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:10:58 vm11.local ceph-mon[53973]: pgmap v3146: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:10:59.741 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:10:59.741 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:10:59.768 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:10:59.769 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:00.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:00 vm10.local ceph-mon[53712]: pgmap v3147: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:00 vm11.local ceph-mon[53973]: pgmap v3147: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:02.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:02 vm10.local ceph-mon[53712]: pgmap v3148: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:02 vm11.local ceph-mon[53973]: pgmap v3148: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:04.770 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:04.771 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:04.800 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:04.801 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:04.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:04 vm10.local ceph-mon[53712]: pgmap v3149: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:04 vm11.local ceph-mon[53973]: pgmap v3149: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:06.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:06 vm10.local ceph-mon[53712]: pgmap v3150: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:06 vm11.local ceph-mon[53973]: pgmap v3150: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:08.938 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:08 vm10.local ceph-mon[53712]: pgmap v3151: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:08.992 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:08 vm11.local ceph-mon[53973]: pgmap v3151: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:09.802 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:09.802 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:09.830 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:09.831 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:10 vm11.local ceph-mon[53973]: pgmap v3152: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:10 vm10.local ceph-mon[53712]: pgmap v3152: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:12 vm11.local ceph-mon[53973]: pgmap v3153: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:12 vm10.local ceph-mon[53712]: pgmap v3153: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:14.832 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:14.833 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:14.860 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:14.861 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:14 vm11.local ceph-mon[53973]: pgmap v3154: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:14 vm10.local ceph-mon[53712]: pgmap v3154: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:17.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:16 vm11.local ceph-mon[53973]: pgmap v3155: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:16 vm10.local ceph-mon[53712]: pgmap v3155: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:18 vm11.local ceph-mon[53973]: pgmap v3156: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:19.140 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:18 vm10.local ceph-mon[53712]: pgmap v3156: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:19.862 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:19.863 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:19.891 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:19.892 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:20 vm11.local ceph-mon[53973]: pgmap v3157: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:20 vm10.local ceph-mon[53712]: pgmap v3157: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:22 vm11.local ceph-mon[53973]: pgmap v3158: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:23.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:22 vm10.local ceph-mon[53712]: pgmap v3158: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:24.893 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:24.894 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:24.922 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:24.922 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:24 vm11.local ceph-mon[53973]: pgmap v3159: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:25.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:24 vm10.local ceph-mon[53712]: pgmap v3159: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:26 vm11.local ceph-mon[53973]: pgmap v3160: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:26 vm10.local ceph-mon[53712]: pgmap v3160: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:28 vm11.local ceph-mon[53973]: pgmap v3161: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:11:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:11:29.140 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:28 vm10.local ceph-mon[53712]: pgmap v3161: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:29.141 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:11:29.141 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:11:29.924 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:29.924 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:29.953 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:29.953 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:31.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:30 vm11.local ceph-mon[53973]: pgmap v3162: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:31.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:30 vm10.local ceph-mon[53712]: pgmap v3162: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:33.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:32 vm11.local ceph-mon[53973]: pgmap v3163: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:33.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:32 vm10.local ceph-mon[53712]: pgmap v3163: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:34.955 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:34.955 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:34.983 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:34.984 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:34 vm11.local ceph-mon[53973]: pgmap v3164: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:34 vm10.local ceph-mon[53712]: pgmap v3164: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:37.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:36 vm11.local ceph-mon[53973]: pgmap v3165: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:37.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:36 vm10.local ceph-mon[53712]: pgmap v3165: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:38 vm11.local ceph-mon[53973]: pgmap v3166: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:39.142 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:38 vm10.local ceph-mon[53712]: pgmap v3166: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:39.985 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:39.986 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:40.015 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:40.016 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:40 vm11.local ceph-mon[53973]: pgmap v3167: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:40 vm10.local ceph-mon[53712]: pgmap v3167: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:43.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:42 vm11.local ceph-mon[53973]: pgmap v3168: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:42 vm10.local ceph-mon[53712]: pgmap v3168: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:43.826 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:11:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:11:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:11:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:11:44.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:43 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:11:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:11:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:11:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:11:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:11:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:43 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:11:45.018 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:45.018 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:45.045 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:45.046 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:44 vm11.local ceph-mon[53973]: pgmap v3169: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:44 vm10.local ceph-mon[53712]: pgmap v3169: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:46 vm11.local ceph-mon[53973]: pgmap v3170: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:46 vm10.local ceph-mon[53712]: pgmap v3170: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:48 vm11.local ceph-mon[53973]: pgmap v3171: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:49.142 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:48 vm10.local ceph-mon[53712]: pgmap v3171: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:50.047 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:50.048 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:50.086 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:50.087 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:50 vm10.local ceph-mon[53712]: pgmap v3172: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:50 vm11.local ceph-mon[53973]: pgmap v3172: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:52 vm10.local ceph-mon[53712]: pgmap v3173: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:52 vm11.local ceph-mon[53973]: pgmap v3173: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:55.089 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:11:55.090 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:11:55.119 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:11:55.119 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:11:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:54 vm10.local ceph-mon[53712]: pgmap v3174: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:54 vm11.local ceph-mon[53973]: pgmap v3174: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:56 vm10.local ceph-mon[53712]: pgmap v3175: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:56 vm11.local ceph-mon[53973]: pgmap v3175: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:11:59.143 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:11:58 vm10.local ceph-mon[53712]: pgmap v3176: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:11:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:11:58 vm11.local ceph-mon[53973]: pgmap v3176: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:00.121 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:00.122 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:00.150 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:00.150 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:00 vm10.local ceph-mon[53712]: pgmap v3177: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:00 vm11.local ceph-mon[53973]: pgmap v3177: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:02 vm10.local ceph-mon[53712]: pgmap v3178: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:02 vm11.local ceph-mon[53973]: pgmap v3178: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:05.152 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:05.152 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:05.178 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:05.179 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:04 vm10.local ceph-mon[53712]: pgmap v3179: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:04 vm11.local ceph-mon[53973]: pgmap v3179: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:06 vm10.local ceph-mon[53712]: pgmap v3180: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:06 vm11.local ceph-mon[53973]: pgmap v3180: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:09.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:08 vm10.local ceph-mon[53712]: pgmap v3181: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:08 vm11.local ceph-mon[53973]: pgmap v3181: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:10.181 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:10.181 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:10.208 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:10.209 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:10 vm10.local ceph-mon[53712]: pgmap v3182: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:10 vm11.local ceph-mon[53973]: pgmap v3182: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:12 vm10.local ceph-mon[53712]: pgmap v3183: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:12 vm11.local ceph-mon[53973]: pgmap v3183: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:14 vm10.local ceph-mon[53712]: pgmap v3184: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:15.210 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:15.211 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:15.238 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:15.238 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:14 vm11.local ceph-mon[53973]: pgmap v3184: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:16 vm10.local ceph-mon[53712]: pgmap v3185: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:16 vm11.local ceph-mon[53973]: pgmap v3185: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:19.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:18 vm10.local ceph-mon[53712]: pgmap v3186: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:18 vm11.local ceph-mon[53973]: pgmap v3186: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:20.240 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:20.240 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:20.267 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:20.267 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:20 vm11.local ceph-mon[53973]: pgmap v3187: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:20 vm10.local ceph-mon[53712]: pgmap v3187: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:22 vm11.local ceph-mon[53973]: pgmap v3188: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:22 vm10.local ceph-mon[53712]: pgmap v3188: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:25.269 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:25.269 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:25.295 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:25.296 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:24 vm11.local ceph-mon[53973]: pgmap v3189: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:24 vm10.local ceph-mon[53712]: pgmap v3189: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:26 vm11.local ceph-mon[53973]: pgmap v3190: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:26 vm10.local ceph-mon[53712]: pgmap v3190: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:28 vm11.local ceph-mon[53973]: pgmap v3191: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:12:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:12:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:28 vm10.local ceph-mon[53712]: pgmap v3191: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:12:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:12:30.297 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:30.297 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:30.324 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:30.324 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:30 vm11.local ceph-mon[53973]: pgmap v3192: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:30 vm10.local ceph-mon[53712]: pgmap v3192: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:32 vm11.local ceph-mon[53973]: pgmap v3193: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:32 vm10.local ceph-mon[53712]: pgmap v3193: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:35.326 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:35.326 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:34 vm11.local ceph-mon[53973]: pgmap v3194: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:35.357 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:35.358 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:34 vm10.local ceph-mon[53712]: pgmap v3194: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:36 vm11.local ceph-mon[53973]: pgmap v3195: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:36 vm10.local ceph-mon[53712]: pgmap v3195: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:38 vm11.local ceph-mon[53973]: pgmap v3196: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:38 vm10.local ceph-mon[53712]: pgmap v3196: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:40.359 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:40.360 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:40.399 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:40.399 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:40 vm11.local ceph-mon[53973]: pgmap v3197: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:40 vm10.local ceph-mon[53712]: pgmap v3197: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:42 vm11.local ceph-mon[53973]: pgmap v3198: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:43.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:42 vm10.local ceph-mon[53712]: pgmap v3198: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:44.003 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:12:44.003 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:12:44.003 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:12:44.003 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:12:44.003 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:12:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:12:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:12:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:12:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:12:44.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:12:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:45 vm11.local ceph-mon[53973]: pgmap v3199: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:45.401 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:45.401 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:45.427 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:45.428 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:45 vm10.local ceph-mon[53712]: pgmap v3199: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:47 vm11.local ceph-mon[53973]: pgmap v3200: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:47 vm10.local ceph-mon[53712]: pgmap v3200: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:49.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:49 vm10.local ceph-mon[53712]: pgmap v3201: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:49 vm11.local ceph-mon[53973]: pgmap v3201: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:50.429 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:50.430 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:50.456 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:50.456 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:51 vm11.local ceph-mon[53973]: pgmap v3202: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:51 vm10.local ceph-mon[53712]: pgmap v3202: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:53 vm11.local ceph-mon[53973]: pgmap v3203: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:53 vm10.local ceph-mon[53712]: pgmap v3203: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:55 vm11.local ceph-mon[53973]: pgmap v3204: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:55.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:55 vm10.local ceph-mon[53712]: pgmap v3204: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:55.458 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:12:55.458 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:12:55.483 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:12:55.484 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:12:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:57 vm11.local ceph-mon[53973]: pgmap v3205: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:57 vm10.local ceph-mon[53712]: pgmap v3205: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:12:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:12:59 vm11.local ceph-mon[53973]: pgmap v3206: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:12:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:12:59 vm10.local ceph-mon[53712]: pgmap v3206: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:00.485 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:00.486 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:00.512 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:00.513 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:01 vm11.local ceph-mon[53973]: pgmap v3207: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:01 vm10.local ceph-mon[53712]: pgmap v3207: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:03 vm11.local ceph-mon[53973]: pgmap v3208: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:03 vm10.local ceph-mon[53712]: pgmap v3208: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:05 vm11.local ceph-mon[53973]: pgmap v3209: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:05 vm10.local ceph-mon[53712]: pgmap v3209: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:05.514 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:05.515 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:05.540 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:05.541 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:07 vm11.local ceph-mon[53973]: pgmap v3210: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:07 vm10.local ceph-mon[53712]: pgmap v3210: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:09 vm11.local ceph-mon[53973]: pgmap v3211: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:09.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:09 vm10.local ceph-mon[53712]: pgmap v3211: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:10.542 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:10.542 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:10.569 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:10.570 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:11 vm11.local ceph-mon[53973]: pgmap v3212: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:11 vm10.local ceph-mon[53712]: pgmap v3212: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:13 vm11.local ceph-mon[53973]: pgmap v3213: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:13 vm10.local ceph-mon[53712]: pgmap v3213: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:15 vm11.local ceph-mon[53973]: pgmap v3214: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:15 vm10.local ceph-mon[53712]: pgmap v3214: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:15.571 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:15.572 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:15.598 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:15.598 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:17 vm11.local ceph-mon[53973]: pgmap v3215: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:17 vm10.local ceph-mon[53712]: pgmap v3215: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:19.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:19 vm10.local ceph-mon[53712]: pgmap v3216: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:19 vm11.local ceph-mon[53973]: pgmap v3216: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:20.600 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:20.600 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:20.629 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:20.630 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:21 vm10.local ceph-mon[53712]: pgmap v3217: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:21 vm11.local ceph-mon[53973]: pgmap v3217: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:23 vm10.local ceph-mon[53712]: pgmap v3218: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:23 vm11.local ceph-mon[53973]: pgmap v3218: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:25 vm10.local ceph-mon[53712]: pgmap v3219: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:25 vm11.local ceph-mon[53973]: pgmap v3219: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:25.631 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:25.631 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:25.659 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:25.660 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:27 vm10.local ceph-mon[53712]: pgmap v3220: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:27 vm11.local ceph-mon[53973]: pgmap v3220: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:29 vm10.local ceph-mon[53712]: pgmap v3221: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:13:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:13:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:29 vm11.local ceph-mon[53973]: pgmap v3221: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:13:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:13:30.661 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:30.662 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:30.687 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:30.688 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:31 vm10.local ceph-mon[53712]: pgmap v3222: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:31 vm11.local ceph-mon[53973]: pgmap v3222: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:33 vm10.local ceph-mon[53712]: pgmap v3223: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:33 vm11.local ceph-mon[53973]: pgmap v3223: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:35 vm10.local ceph-mon[53712]: pgmap v3224: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:35 vm11.local ceph-mon[53973]: pgmap v3224: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 341 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:35.689 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:35.689 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:35.716 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:35.716 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:37 vm10.local ceph-mon[53712]: pgmap v3225: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:37 vm11.local ceph-mon[53973]: pgmap v3225: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:39 vm10.local ceph-mon[53712]: pgmap v3226: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:39 vm11.local ceph-mon[53973]: pgmap v3226: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:40.718 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:40.718 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:40.744 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:40.745 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:41 vm10.local ceph-mon[53712]: pgmap v3227: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:41 vm11.local ceph-mon[53973]: pgmap v3227: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:43 vm10.local ceph-mon[53712]: pgmap v3228: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:43 vm11.local ceph-mon[53973]: pgmap v3228: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:13:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:13:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:13:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:13:44.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:13:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:13:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:13:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:13:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:13:44.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:13:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:45 vm10.local ceph-mon[53712]: pgmap v3229: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:45 vm11.local ceph-mon[53973]: pgmap v3229: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:45.746 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:45.746 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:45.773 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:45.773 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:47 vm10.local ceph-mon[53712]: pgmap v3230: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:47 vm11.local ceph-mon[53973]: pgmap v3230: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:49.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:49 vm10.local ceph-mon[53712]: pgmap v3231: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:49 vm11.local ceph-mon[53973]: pgmap v3231: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:50.775 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:50.775 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:50.803 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:50.804 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:51 vm10.local ceph-mon[53712]: pgmap v3232: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:51 vm11.local ceph-mon[53973]: pgmap v3232: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:53 vm11.local ceph-mon[53973]: pgmap v3233: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:53.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:53 vm10.local ceph-mon[53712]: pgmap v3233: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:55 vm11.local ceph-mon[53973]: pgmap v3234: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:55.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:55 vm10.local ceph-mon[53712]: pgmap v3234: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:55.805 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:13:55.806 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:13:55.832 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:13:55.832 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:13:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:57 vm11.local ceph-mon[53973]: pgmap v3235: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:57.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:57 vm10.local ceph-mon[53712]: pgmap v3235: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:13:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:13:59 vm11.local ceph-mon[53973]: pgmap v3236: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:13:59.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:13:59 vm10.local ceph-mon[53712]: pgmap v3236: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:00.833 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:00.834 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:00.861 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:00.861 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:01.464 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:01 vm10.local ceph-mon[53712]: pgmap v3237: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:01 vm11.local ceph-mon[53973]: pgmap v3237: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:03 vm11.local ceph-mon[53973]: pgmap v3238: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:03.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:03 vm10.local ceph-mon[53712]: pgmap v3238: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:05 vm11.local ceph-mon[53973]: pgmap v3239: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:05.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:05 vm10.local ceph-mon[53712]: pgmap v3239: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:05.863 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:05.863 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:05.890 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:05.890 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:06.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:06 vm11.local ceph-mon[53973]: pgmap v3240: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:06.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:06 vm10.local ceph-mon[53712]: pgmap v3240: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:08.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:08 vm10.local ceph-mon[53712]: pgmap v3241: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:08 vm11.local ceph-mon[53973]: pgmap v3241: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:10.892 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:10.892 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:10.921 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:10.921 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:10.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:10 vm10.local ceph-mon[53712]: pgmap v3242: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:10 vm11.local ceph-mon[53973]: pgmap v3242: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:12.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:12 vm10.local ceph-mon[53712]: pgmap v3243: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:12 vm11.local ceph-mon[53973]: pgmap v3243: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:14.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:14 vm10.local ceph-mon[53712]: pgmap v3244: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:15.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:14 vm11.local ceph-mon[53973]: pgmap v3244: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:15.923 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:15.923 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:15.951 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:15.952 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:16.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:16 vm10.local ceph-mon[53712]: pgmap v3245: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:17.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:16 vm11.local ceph-mon[53973]: pgmap v3245: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:18.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:18 vm10.local ceph-mon[53712]: pgmap v3246: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:19.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:18 vm11.local ceph-mon[53973]: pgmap v3246: 97 pgs: 97 active+clean; 453 KiB data, 83 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:20.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:20 vm10.local ceph-mon[53712]: pgmap v3247: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:20.953 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:20.953 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:20.979 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:20.980 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:21.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:20 vm11.local ceph-mon[53973]: pgmap v3247: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:22.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:22 vm10.local ceph-mon[53712]: pgmap v3248: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:23.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:22 vm11.local ceph-mon[53973]: pgmap v3248: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:24.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:24 vm10.local ceph-mon[53712]: pgmap v3249: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:25.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:24 vm11.local ceph-mon[53973]: pgmap v3249: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:25.981 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:25.982 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:26.008 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:26.008 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:26 vm10.local ceph-mon[53712]: pgmap v3250: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:26 vm11.local ceph-mon[53973]: pgmap v3250: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:28.938 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:28 vm10.local ceph-mon[53712]: pgmap v3251: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:28.938 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:14:28.938 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:14:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:28 vm11.local ceph-mon[53973]: pgmap v3251: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:14:29.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:14:31.010 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:31.010 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:31.036 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:31.037 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:31.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:30 vm11.local ceph-mon[53973]: pgmap v3252: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:31.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:30 vm10.local ceph-mon[53712]: pgmap v3252: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:33.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:32 vm11.local ceph-mon[53973]: pgmap v3253: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:33.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:32 vm10.local ceph-mon[53712]: pgmap v3253: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:35.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:34 vm11.local ceph-mon[53973]: pgmap v3254: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:35.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:34 vm10.local ceph-mon[53712]: pgmap v3254: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:36.038 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:36.039 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:36.064 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:36.065 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:37.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:36 vm11.local ceph-mon[53973]: pgmap v3255: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:37.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:36 vm10.local ceph-mon[53712]: pgmap v3255: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:38.991 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:38 vm11.local ceph-mon[53973]: pgmap v3256: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:39.157 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:38 vm10.local ceph-mon[53712]: pgmap v3256: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:41.066 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:41.066 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:40 vm11.local ceph-mon[53973]: pgmap v3257: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:41.092 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:41.093 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:40 vm10.local ceph-mon[53712]: pgmap v3257: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:43.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:42 vm11.local ceph-mon[53973]: pgmap v3258: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:42 vm10.local ceph-mon[53712]: pgmap v3258: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:44 vm11.local ceph-mon[53973]: pgmap v3259: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:14:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:14:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:14:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:14:45.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:44 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:14:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:44 vm10.local ceph-mon[53712]: pgmap v3259: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:14:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:14:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:14:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:14:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:44 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:14:46.094 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:46.095 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:46.120 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:46.121 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:47.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:46 vm11.local ceph-mon[53973]: pgmap v3260: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:46 vm10.local ceph-mon[53712]: pgmap v3260: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:49.158 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:48 vm10.local ceph-mon[53712]: pgmap v3261: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:48 vm11.local ceph-mon[53973]: pgmap v3261: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:51.122 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:51.123 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:51.150 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:51.150 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:50 vm10.local ceph-mon[53712]: pgmap v3262: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:50 vm11.local ceph-mon[53973]: pgmap v3262: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:52 vm10.local ceph-mon[53712]: pgmap v3263: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:52 vm11.local ceph-mon[53973]: pgmap v3263: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:54 vm10.local ceph-mon[53712]: pgmap v3264: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:54 vm11.local ceph-mon[53973]: pgmap v3264: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:56.152 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:14:56.152 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:14:56.180 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:14:56.180 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:14:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:56 vm10.local ceph-mon[53712]: pgmap v3265: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:56 vm11.local ceph-mon[53973]: pgmap v3265: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:14:59.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:14:58 vm10.local ceph-mon[53712]: pgmap v3266: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:14:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:14:58 vm11.local ceph-mon[53973]: pgmap v3266: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:01.182 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:01.182 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:01.185 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:00 vm10.local ceph-mon[53712]: pgmap v3267: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:01.208 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:01.209 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:00 vm11.local ceph-mon[53973]: pgmap v3267: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:02 vm10.local ceph-mon[53712]: pgmap v3268: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:02 vm11.local ceph-mon[53973]: pgmap v3268: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:04 vm10.local ceph-mon[53712]: pgmap v3269: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:04 vm11.local ceph-mon[53973]: pgmap v3269: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:06.210 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:06.211 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:06.238 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:06.238 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:06 vm10.local ceph-mon[53712]: pgmap v3270: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:06 vm11.local ceph-mon[53973]: pgmap v3270: 97 pgs: 97 active+clean; 453 KiB data, 87 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:08 vm11.local ceph-mon[53973]: pgmap v3271: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:08 vm10.local ceph-mon[53712]: pgmap v3271: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:11.240 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:11.240 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:11.266 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:11.267 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:10 vm11.local ceph-mon[53973]: pgmap v3272: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:10 vm10.local ceph-mon[53712]: pgmap v3272: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:12 vm11.local ceph-mon[53973]: pgmap v3273: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:12 vm10.local ceph-mon[53712]: pgmap v3273: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:14 vm11.local ceph-mon[53973]: pgmap v3274: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:14 vm10.local ceph-mon[53712]: pgmap v3274: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:16.268 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:16.269 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:16.295 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:16.296 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:16 vm11.local ceph-mon[53973]: pgmap v3275: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:16 vm10.local ceph-mon[53712]: pgmap v3275: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:18 vm10.local ceph-mon[53712]: pgmap v3276: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:18 vm11.local ceph-mon[53973]: pgmap v3276: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:21.297 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:21.297 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:21.324 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:21.324 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:20 vm11.local ceph-mon[53973]: pgmap v3277: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:20 vm10.local ceph-mon[53712]: pgmap v3277: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:22 vm11.local ceph-mon[53973]: pgmap v3278: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:22 vm10.local ceph-mon[53712]: pgmap v3278: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:24 vm11.local ceph-mon[53973]: pgmap v3279: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:24 vm10.local ceph-mon[53712]: pgmap v3279: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:26.325 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:26.326 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:26.353 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:26.353 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:26 vm11.local ceph-mon[53973]: pgmap v3280: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:26 vm10.local ceph-mon[53712]: pgmap v3280: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:28 vm11.local ceph-mon[53973]: pgmap v3281: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:15:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:15:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:28 vm10.local ceph-mon[53712]: pgmap v3281: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:15:29.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:15:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:30 vm11.local ceph-mon[53973]: pgmap v3282: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:31.354 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:31.355 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:31.381 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:31.381 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:30 vm10.local ceph-mon[53712]: pgmap v3282: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:32 vm11.local ceph-mon[53973]: pgmap v3283: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:32 vm10.local ceph-mon[53712]: pgmap v3283: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:35 vm11.local ceph-mon[53973]: pgmap v3284: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:35 vm10.local ceph-mon[53712]: pgmap v3284: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:36.383 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:36.383 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:36.409 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:36.409 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:37 vm11.local ceph-mon[53973]: pgmap v3285: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:37 vm10.local ceph-mon[53712]: pgmap v3285: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:39 vm11.local ceph-mon[53973]: pgmap v3286: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:39.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:39 vm10.local ceph-mon[53712]: pgmap v3286: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:41.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:41 vm11.local ceph-mon[53973]: pgmap v3287: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:41.410 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:41.411 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:41 vm10.local ceph-mon[53712]: pgmap v3287: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:41.436 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:41.436 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:43 vm11.local ceph-mon[53973]: pgmap v3288: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:43 vm10.local ceph-mon[53712]: pgmap v3288: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:45.024 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:45 vm11.local ceph-mon[53973]: pgmap v3289: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:45.024 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:15:45.024 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:15:45.024 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:15:45.024 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:15:45.024 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:15:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:45 vm10.local ceph-mon[53712]: pgmap v3289: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:15:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:15:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:15:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:15:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:15:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:15:46.437 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:46.438 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:46.464 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:46.464 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:47 vm11.local ceph-mon[53973]: pgmap v3290: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:47 vm10.local ceph-mon[53712]: pgmap v3290: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:49.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:49 vm10.local ceph-mon[53712]: pgmap v3291: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:49 vm11.local ceph-mon[53973]: pgmap v3291: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:51.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:51 vm11.local ceph-mon[53973]: pgmap v3292: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:15:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:51 vm10.local ceph-mon[53712]: pgmap v3292: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:15:51.465 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:51.466 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:51.491 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:51.492 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:53 vm11.local ceph-mon[53973]: pgmap v3293: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:15:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:53 vm10.local ceph-mon[53712]: pgmap v3293: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:15:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:55 vm11.local ceph-mon[53973]: pgmap v3294: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:55 vm10.local ceph-mon[53712]: pgmap v3294: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:56.493 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:15:56.493 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:15:56.518 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:15:56.519 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:15:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:57 vm11.local ceph-mon[53973]: pgmap v3295: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:15:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:57 vm10.local ceph-mon[53712]: pgmap v3295: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 85 B/s wr, 0 op/s 2026-03-09T19:15:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:15:59 vm11.local ceph-mon[53973]: pgmap v3296: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:15:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:15:59 vm10.local ceph-mon[53712]: pgmap v3296: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:01 vm11.local ceph-mon[53973]: pgmap v3297: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:01 vm10.local ceph-mon[53712]: pgmap v3297: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:01.520 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:01.521 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:01.546 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:01.546 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:03 vm11.local ceph-mon[53973]: pgmap v3298: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:03 vm10.local ceph-mon[53712]: pgmap v3298: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:05 vm11.local ceph-mon[53973]: pgmap v3299: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:05 vm10.local ceph-mon[53712]: pgmap v3299: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:06.548 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:06.548 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:06.574 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:06.574 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:07 vm11.local ceph-mon[53973]: pgmap v3300: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:07 vm10.local ceph-mon[53712]: pgmap v3300: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:09 vm11.local ceph-mon[53973]: pgmap v3301: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:09 vm10.local ceph-mon[53712]: pgmap v3301: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:11 vm10.local ceph-mon[53712]: pgmap v3302: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:11.576 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:11.576 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:11 vm11.local ceph-mon[53973]: pgmap v3302: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:11.603 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:11.604 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:13 vm10.local ceph-mon[53712]: pgmap v3303: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:13 vm11.local ceph-mon[53973]: pgmap v3303: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:15 vm10.local ceph-mon[53712]: pgmap v3304: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:15 vm11.local ceph-mon[53973]: pgmap v3304: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:16.605 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:16.606 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:16.633 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:16.634 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:17 vm10.local ceph-mon[53712]: pgmap v3305: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:17 vm11.local ceph-mon[53973]: pgmap v3305: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:19 vm10.local ceph-mon[53712]: pgmap v3306: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:19 vm11.local ceph-mon[53973]: pgmap v3306: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:21 vm10.local ceph-mon[53712]: pgmap v3307: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:21 vm11.local ceph-mon[53973]: pgmap v3307: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:21.635 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:21.635 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:21.662 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:21.662 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:23 vm10.local ceph-mon[53712]: pgmap v3308: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:23 vm11.local ceph-mon[53973]: pgmap v3308: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:25 vm10.local ceph-mon[53712]: pgmap v3309: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:25 vm11.local ceph-mon[53973]: pgmap v3309: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:26.664 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:26.664 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:26.691 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:26.692 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:27 vm10.local ceph-mon[53712]: pgmap v3310: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:27 vm11.local ceph-mon[53973]: pgmap v3310: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:29 vm10.local ceph-mon[53712]: pgmap v3311: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:16:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:16:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:29 vm11.local ceph-mon[53973]: pgmap v3311: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:16:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:16:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:31 vm10.local ceph-mon[53712]: pgmap v3312: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:31 vm11.local ceph-mon[53973]: pgmap v3312: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:31.693 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:31.693 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:31.720 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:31.720 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:33 vm10.local ceph-mon[53712]: pgmap v3313: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:33 vm11.local ceph-mon[53973]: pgmap v3313: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:35 vm11.local ceph-mon[53973]: pgmap v3314: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:35 vm10.local ceph-mon[53712]: pgmap v3314: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:36.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:36 vm11.local ceph-mon[53973]: pgmap v3315: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:36.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:36 vm10.local ceph-mon[53712]: pgmap v3315: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:36.721 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:36.722 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:36.748 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:36.749 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:38 vm11.local ceph-mon[53973]: pgmap v3316: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:39.166 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:38 vm10.local ceph-mon[53712]: pgmap v3316: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:41.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:40 vm11.local ceph-mon[53973]: pgmap v3317: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:41.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:40 vm10.local ceph-mon[53712]: pgmap v3317: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:41.750 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:41.750 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:41.776 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:41.777 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:43.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:42 vm11.local ceph-mon[53973]: pgmap v3318: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:42 vm10.local ceph-mon[53712]: pgmap v3318: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:45.061 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:44 vm10.local ceph-mon[53712]: pgmap v3319: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:45.063 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:44 vm11.local ceph-mon[53973]: pgmap v3319: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:16:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:16:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:16:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:16:46.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:16:46.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:16:46.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:16:46.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:16:46.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:16:46.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:16:46.778 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:46.778 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:46.809 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:46.809 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:47 vm11.local ceph-mon[53973]: pgmap v3320: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:47 vm10.local ceph-mon[53712]: pgmap v3320: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:49.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:49 vm10.local ceph-mon[53712]: pgmap v3321: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:49 vm11.local ceph-mon[53973]: pgmap v3321: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:51.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:51 vm10.local ceph-mon[53712]: pgmap v3322: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:51 vm11.local ceph-mon[53973]: pgmap v3322: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:51.811 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:51.811 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:51.838 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:51.838 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:53.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:53 vm10.local ceph-mon[53712]: pgmap v3323: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:53.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:53 vm11.local ceph-mon[53973]: pgmap v3323: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:55.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:55 vm10.local ceph-mon[53712]: pgmap v3324: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:55.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:55 vm11.local ceph-mon[53973]: pgmap v3324: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:56.839 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:16:56.840 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:16:56.964 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:16:56.964 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:16:57.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:57 vm10.local ceph-mon[53712]: pgmap v3325: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:57.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:57 vm11.local ceph-mon[53973]: pgmap v3325: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:16:59.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:16:59 vm10.local ceph-mon[53712]: pgmap v3326: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:16:59.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:16:59 vm11.local ceph-mon[53973]: pgmap v3326: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:01 vm10.local ceph-mon[53712]: pgmap v3327: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:01.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:01 vm11.local ceph-mon[53973]: pgmap v3327: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:01.966 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:01.966 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:01.992 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:01.992 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:03 vm10.local ceph-mon[53712]: pgmap v3328: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:03.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:03 vm11.local ceph-mon[53973]: pgmap v3328: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:05 vm10.local ceph-mon[53712]: pgmap v3329: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:05.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:05 vm11.local ceph-mon[53973]: pgmap v3329: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:06.994 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:06.994 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:07.020 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:07.020 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:07 vm10.local ceph-mon[53712]: pgmap v3330: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:07.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:07 vm11.local ceph-mon[53973]: pgmap v3330: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:09 vm10.local ceph-mon[53712]: pgmap v3331: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:09 vm11.local ceph-mon[53973]: pgmap v3331: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:11 vm11.local ceph-mon[53973]: pgmap v3332: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:11.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:11 vm10.local ceph-mon[53712]: pgmap v3332: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:12.021 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:12.022 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:12.149 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:12.149 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:12 vm11.local ceph-mon[53973]: pgmap v3333: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:12 vm10.local ceph-mon[53712]: pgmap v3333: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:14 vm10.local ceph-mon[53712]: pgmap v3334: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:14 vm11.local ceph-mon[53973]: pgmap v3334: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:17.151 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:17.151 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:17.178 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:17.179 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:16 vm11.local ceph-mon[53973]: pgmap v3335: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:16 vm10.local ceph-mon[53712]: pgmap v3335: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:19.237 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:18 vm10.local ceph-mon[53712]: pgmap v3336: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:18 vm11.local ceph-mon[53973]: pgmap v3336: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:20 vm11.local ceph-mon[53973]: pgmap v3337: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:20 vm10.local ceph-mon[53712]: pgmap v3337: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:22.180 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:22.181 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:22.206 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:22.207 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:22 vm11.local ceph-mon[53973]: pgmap v3338: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:22 vm10.local ceph-mon[53712]: pgmap v3338: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:25 vm11.local ceph-mon[53973]: pgmap v3339: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:25.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:25 vm10.local ceph-mon[53712]: pgmap v3339: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:27.208 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:27.209 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:27 vm11.local ceph-mon[53973]: pgmap v3340: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:27 vm10.local ceph-mon[53712]: pgmap v3340: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:27.477 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:27.479 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:29 vm11.local ceph-mon[53973]: pgmap v3341: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:17:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:17:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:29 vm10.local ceph-mon[53712]: pgmap v3341: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:17:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:17:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:31 vm11.local ceph-mon[53973]: pgmap v3342: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:31 vm10.local ceph-mon[53712]: pgmap v3342: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:32.480 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:32.480 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:32.508 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:32.509 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:33 vm11.local ceph-mon[53973]: pgmap v3343: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:33 vm10.local ceph-mon[53712]: pgmap v3343: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:35 vm10.local ceph-mon[53712]: pgmap v3344: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:35 vm11.local ceph-mon[53973]: pgmap v3344: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:37.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:37 vm10.local ceph-mon[53712]: pgmap v3345: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:37.510 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:37.511 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:37.538 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:37.539 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:37 vm11.local ceph-mon[53973]: pgmap v3345: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:39 vm10.local ceph-mon[53712]: pgmap v3346: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:39 vm11.local ceph-mon[53973]: pgmap v3346: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:41 vm10.local ceph-mon[53712]: pgmap v3347: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:41 vm11.local ceph-mon[53973]: pgmap v3347: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:42.540 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:42.541 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:42.567 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:42.567 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:43 vm10.local ceph-mon[53712]: pgmap v3348: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:43 vm11.local ceph-mon[53973]: pgmap v3348: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:45.360 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:45 vm10.local ceph-mon[53712]: pgmap v3349: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:45.360 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:17:45.360 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:17:45.361 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:45 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:17:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:45 vm11.local ceph-mon[53973]: pgmap v3349: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:17:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:17:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:45 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:17:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:17:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:46 vm11.local ceph-mon[53973]: pgmap v3350: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:17:46.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:17:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:17:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:46 vm10.local ceph-mon[53712]: pgmap v3350: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:17:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:17:47.569 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:47.569 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:47.595 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:47.596 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:49.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:48 vm11.local ceph-mon[53973]: pgmap v3351: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:49.171 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:48 vm10.local ceph-mon[53712]: pgmap v3351: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:51.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:50 vm11.local ceph-mon[53973]: pgmap v3352: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:51.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:50 vm10.local ceph-mon[53712]: pgmap v3352: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:52.598 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:52.598 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:52.625 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:52.626 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:52 vm11.local ceph-mon[53973]: pgmap v3353: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:52 vm10.local ceph-mon[53712]: pgmap v3353: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:54 vm11.local ceph-mon[53973]: pgmap v3354: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:54 vm10.local ceph-mon[53712]: pgmap v3354: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:56 vm10.local ceph-mon[53712]: pgmap v3355: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:56 vm11.local ceph-mon[53973]: pgmap v3355: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:17:57.627 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:17:57.628 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:17:57.656 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:17:57.657 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:17:59.171 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:17:58 vm10.local ceph-mon[53712]: pgmap v3356: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:17:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:17:58 vm11.local ceph-mon[53973]: pgmap v3356: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:00 vm11.local ceph-mon[53973]: pgmap v3357: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:00 vm10.local ceph-mon[53712]: pgmap v3357: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:02.659 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:02.660 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:02.691 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:02.692 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:03 vm11.local ceph-mon[53973]: pgmap v3358: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:03 vm10.local ceph-mon[53712]: pgmap v3358: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:05 vm11.local ceph-mon[53973]: pgmap v3359: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:05 vm10.local ceph-mon[53712]: pgmap v3359: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:07 vm11.local ceph-mon[53973]: pgmap v3360: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:07 vm10.local ceph-mon[53712]: pgmap v3360: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:07.694 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:07.694 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:07.722 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:07.723 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:09.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:09 vm10.local ceph-mon[53712]: pgmap v3361: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:09.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:09 vm11.local ceph-mon[53973]: pgmap v3361: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:11 vm10.local ceph-mon[53712]: pgmap v3362: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:11.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:11 vm11.local ceph-mon[53973]: pgmap v3362: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:12.724 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:12.725 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:12.751 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:12.752 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:13 vm10.local ceph-mon[53712]: pgmap v3363: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:13.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:13 vm11.local ceph-mon[53973]: pgmap v3363: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:15 vm10.local ceph-mon[53712]: pgmap v3364: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:15.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:15 vm11.local ceph-mon[53973]: pgmap v3364: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:17 vm10.local ceph-mon[53712]: pgmap v3365: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:17 vm11.local ceph-mon[53973]: pgmap v3365: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:17.753 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:17.753 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:17.830 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:17.830 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:19 vm10.local ceph-mon[53712]: pgmap v3366: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:19 vm11.local ceph-mon[53973]: pgmap v3366: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:21 vm10.local ceph-mon[53712]: pgmap v3367: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:21 vm11.local ceph-mon[53973]: pgmap v3367: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:22.832 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:22.833 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:22.860 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:22.860 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:23 vm10.local ceph-mon[53712]: pgmap v3368: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:23 vm11.local ceph-mon[53973]: pgmap v3368: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:25 vm10.local ceph-mon[53712]: pgmap v3369: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:25 vm11.local ceph-mon[53973]: pgmap v3369: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:27 vm10.local ceph-mon[53712]: pgmap v3370: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:27 vm11.local ceph-mon[53973]: pgmap v3370: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:27.861 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:27.862 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:27.965 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:27.965 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:29 vm10.local ceph-mon[53712]: pgmap v3371: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:18:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:18:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:29 vm11.local ceph-mon[53973]: pgmap v3371: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:18:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:18:31.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:31 vm10.local ceph-mon[53712]: pgmap v3372: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:31 vm11.local ceph-mon[53973]: pgmap v3372: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:32.967 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:32.967 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:32.994 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:32.995 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:33 vm10.local ceph-mon[53712]: pgmap v3373: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:33 vm11.local ceph-mon[53973]: pgmap v3373: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:35 vm11.local ceph-mon[53973]: pgmap v3374: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:35 vm10.local ceph-mon[53712]: pgmap v3374: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:37 vm11.local ceph-mon[53973]: pgmap v3375: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:37 vm10.local ceph-mon[53712]: pgmap v3375: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:37.996 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:37.997 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:38.023 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:38.024 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:39 vm11.local ceph-mon[53973]: pgmap v3376: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:39.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:39 vm10.local ceph-mon[53712]: pgmap v3376: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:41 vm11.local ceph-mon[53973]: pgmap v3377: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:41 vm10.local ceph-mon[53712]: pgmap v3377: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:43.025 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:43.026 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:43.054 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:43.055 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:43 vm11.local ceph-mon[53973]: pgmap v3378: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:43 vm10.local ceph-mon[53712]: pgmap v3378: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:45 vm11.local ceph-mon[53973]: pgmap v3379: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:45.590 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:45 vm10.local ceph-mon[53712]: pgmap v3379: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:46.326 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:18:46.326 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:18:46.326 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:18:46.326 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:18:46.326 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:18:46.326 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:18:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:18:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:18:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:18:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:18:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:18:46.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:18:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:47 vm11.local ceph-mon[53973]: pgmap v3380: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:47.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:47 vm10.local ceph-mon[53712]: pgmap v3380: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:48.056 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:48.057 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:48.083 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:48.084 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:48.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:48 vm11.local ceph-mon[53973]: pgmap v3381: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:48 vm10.local ceph-mon[53712]: pgmap v3381: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:50.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:50 vm11.local ceph-mon[53973]: pgmap v3382: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:50.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:50 vm10.local ceph-mon[53712]: pgmap v3382: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:53.085 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:53.086 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:52 vm11.local ceph-mon[53973]: pgmap v3383: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:53.114 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:53.114 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:52 vm10.local ceph-mon[53712]: pgmap v3383: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:54 vm11.local ceph-mon[53973]: pgmap v3384: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:54 vm10.local ceph-mon[53712]: pgmap v3384: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:56 vm11.local ceph-mon[53973]: pgmap v3385: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:56 vm10.local ceph-mon[53712]: pgmap v3385: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:18:58.115 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:18:58.116 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:18:58.203 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:18:58.204 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:18:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:18:58 vm11.local ceph-mon[53973]: pgmap v3386: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:18:59.175 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:18:58 vm10.local ceph-mon[53712]: pgmap v3386: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:00 vm11.local ceph-mon[53973]: pgmap v3387: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:00 vm10.local ceph-mon[53712]: pgmap v3387: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:02 vm11.local ceph-mon[53973]: pgmap v3388: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:02 vm10.local ceph-mon[53712]: pgmap v3388: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:03.205 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:03.206 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:03.233 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:03.234 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:05.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:04 vm11.local ceph-mon[53973]: pgmap v3389: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:04 vm10.local ceph-mon[53712]: pgmap v3389: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:07.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:06 vm11.local ceph-mon[53973]: pgmap v3390: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:06 vm10.local ceph-mon[53712]: pgmap v3390: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:08.235 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:08.235 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:08.262 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:08.263 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:09.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:08 vm11.local ceph-mon[53973]: pgmap v3391: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:09.175 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:08 vm10.local ceph-mon[53712]: pgmap v3391: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:11.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:10 vm11.local ceph-mon[53973]: pgmap v3392: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:10 vm10.local ceph-mon[53712]: pgmap v3392: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:13.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:12 vm11.local ceph-mon[53973]: pgmap v3393: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:12 vm10.local ceph-mon[53712]: pgmap v3393: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:13.264 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:13.265 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:13.291 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:13.292 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:15 vm11.local ceph-mon[53973]: pgmap v3394: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:14 vm10.local ceph-mon[53712]: pgmap v3394: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:17 vm10.local ceph-mon[53712]: pgmap v3395: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:17.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:17 vm11.local ceph-mon[53973]: pgmap v3395: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:18.293 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:18.294 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:18.322 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:18.323 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:19.236 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:19 vm10.local ceph-mon[53712]: pgmap v3396: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:19.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:19 vm11.local ceph-mon[53973]: pgmap v3396: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:21 vm10.local ceph-mon[53712]: pgmap v3397: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:21.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:21 vm11.local ceph-mon[53973]: pgmap v3397: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:23.324 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:23.325 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:23.351 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:23.351 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:23 vm10.local ceph-mon[53712]: pgmap v3398: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:23 vm11.local ceph-mon[53973]: pgmap v3398: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:25 vm10.local ceph-mon[53712]: pgmap v3399: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:25 vm11.local ceph-mon[53973]: pgmap v3399: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:27.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:27 vm10.local ceph-mon[53712]: pgmap v3400: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:27.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:27 vm11.local ceph-mon[53973]: pgmap v3400: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:28.352 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:28.353 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:28.379 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:28.380 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:29 vm10.local ceph-mon[53712]: pgmap v3401: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:19:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:19:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:29 vm11.local ceph-mon[53973]: pgmap v3401: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:19:29.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:19:31.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:31 vm11.local ceph-mon[53973]: pgmap v3402: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:31.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:31 vm10.local ceph-mon[53712]: pgmap v3402: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:33.381 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:33.382 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:33.411 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:33.412 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:33 vm11.local ceph-mon[53973]: pgmap v3403: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:33 vm10.local ceph-mon[53712]: pgmap v3403: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:35 vm11.local ceph-mon[53973]: pgmap v3404: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:35 vm10.local ceph-mon[53712]: pgmap v3404: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:37 vm11.local ceph-mon[53973]: pgmap v3405: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:37 vm10.local ceph-mon[53712]: pgmap v3405: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:38.414 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:38.414 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:38.442 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:38.442 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:39.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:39 vm11.local ceph-mon[53973]: pgmap v3406: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:39.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:39 vm10.local ceph-mon[53712]: pgmap v3406: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:41 vm11.local ceph-mon[53973]: pgmap v3407: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:41.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:41 vm10.local ceph-mon[53712]: pgmap v3407: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:43.443 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:43.444 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:43.472 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:43.472 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:43.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:43 vm11.local ceph-mon[53973]: pgmap v3408: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:43 vm10.local ceph-mon[53712]: pgmap v3408: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:45.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:45 vm11.local ceph-mon[53973]: pgmap v3409: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:45.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:45 vm10.local ceph-mon[53712]: pgmap v3409: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:46.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:19:46.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:19:46.341 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:19:46.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:19:46.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:19:46.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:19:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:47 vm11.local ceph-mon[53973]: pgmap v3410: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:47 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:19:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:47 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:19:47.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:47 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:19:47.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:47 vm10.local ceph-mon[53712]: pgmap v3410: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:47.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:47 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:19:47.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:47 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:19:47.685 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:47 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:19:48.474 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:48.474 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:48.502 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:48.502 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:49 vm11.local ceph-mon[53973]: pgmap v3411: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:49.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:49 vm10.local ceph-mon[53712]: pgmap v3411: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:51.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:51 vm11.local ceph-mon[53973]: pgmap v3412: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:51.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:51 vm10.local ceph-mon[53712]: pgmap v3412: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:52 vm11.local ceph-mon[53973]: pgmap v3413: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:52 vm10.local ceph-mon[53712]: pgmap v3413: 97 pgs: 97 active+clean; 453 KiB data, 88 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:53.504 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:53.504 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:53.532 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:53.533 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:55.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:54 vm11.local ceph-mon[53973]: pgmap v3414: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:54 vm10.local ceph-mon[53712]: pgmap v3414: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:57.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:56 vm11.local ceph-mon[53973]: pgmap v3415: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:56 vm10.local ceph-mon[53712]: pgmap v3415: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:19:58.535 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:19:58.535 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:19:58.563 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:19:58.564 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:19:59.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:19:58 vm11.local ceph-mon[53973]: pgmap v3416: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:19:59.180 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:19:58 vm10.local ceph-mon[53712]: pgmap v3416: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:00 vm11.local ceph-mon[53973]: pgmap v3417: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:01.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:00 vm11.local ceph-mon[53973]: overall HEALTH_OK 2026-03-09T19:20:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:00 vm10.local ceph-mon[53712]: pgmap v3417: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:01.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:00 vm10.local ceph-mon[53712]: overall HEALTH_OK 2026-03-09T19:20:03.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:02 vm11.local ceph-mon[53973]: pgmap v3418: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:03.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:02 vm10.local ceph-mon[53712]: pgmap v3418: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:03.565 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:03.566 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:03.594 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:03.594 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:05.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:04 vm10.local ceph-mon[53712]: pgmap v3419: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:04 vm11.local ceph-mon[53973]: pgmap v3419: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:07.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:06 vm10.local ceph-mon[53712]: pgmap v3420: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:06 vm11.local ceph-mon[53973]: pgmap v3420: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:08.596 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:08.596 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:08.625 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:08.626 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:09.180 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:08 vm10.local ceph-mon[53712]: pgmap v3421: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:08 vm11.local ceph-mon[53973]: pgmap v3421: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:11.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:10 vm10.local ceph-mon[53712]: pgmap v3422: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:10 vm11.local ceph-mon[53973]: pgmap v3422: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:13.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:12 vm10.local ceph-mon[53712]: pgmap v3423: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:12 vm11.local ceph-mon[53973]: pgmap v3423: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:13.627 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:13.628 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:13.656 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:13.656 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:15.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:14 vm10.local ceph-mon[53712]: pgmap v3424: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:14 vm11.local ceph-mon[53973]: pgmap v3424: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:17.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:16 vm10.local ceph-mon[53712]: pgmap v3425: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:16 vm11.local ceph-mon[53973]: pgmap v3425: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:18.658 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:18.659 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:18.686 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:18.686 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:19.182 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:18 vm10.local ceph-mon[53712]: pgmap v3426: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:18 vm11.local ceph-mon[53973]: pgmap v3426: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:21.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:20 vm10.local ceph-mon[53712]: pgmap v3427: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:20 vm11.local ceph-mon[53973]: pgmap v3427: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:23.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:22 vm11.local ceph-mon[53973]: pgmap v3428: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:22 vm10.local ceph-mon[53712]: pgmap v3428: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:23.688 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:23.688 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:23.718 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:23.718 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:25.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:24 vm11.local ceph-mon[53973]: pgmap v3429: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:25.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:24 vm10.local ceph-mon[53712]: pgmap v3429: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:27.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:26 vm11.local ceph-mon[53973]: pgmap v3430: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:27.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:26 vm10.local ceph-mon[53712]: pgmap v3430: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:28.720 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:28.720 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:28.751 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:28.752 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:28 vm11.local ceph-mon[53973]: pgmap v3431: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:20:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:20:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:28 vm10.local ceph-mon[53712]: pgmap v3431: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:20:29.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:20:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:30 vm11.local ceph-mon[53973]: pgmap v3432: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:30 vm10.local ceph-mon[53712]: pgmap v3432: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:32 vm11.local ceph-mon[53973]: pgmap v3433: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:32 vm10.local ceph-mon[53712]: pgmap v3433: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:33.753 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:33.754 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:33.782 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:33.783 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:34 vm11.local ceph-mon[53973]: pgmap v3434: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:34 vm10.local ceph-mon[53712]: pgmap v3434: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:37.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:36 vm11.local ceph-mon[53973]: pgmap v3435: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:37.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:36 vm10.local ceph-mon[53712]: pgmap v3435: 97 pgs: 97 active+clean; 453 KiB data, 92 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:38.785 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:38.785 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:38.810 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:38.811 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:39.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:38 vm11.local ceph-mon[53973]: pgmap v3436: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:39.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:38 vm10.local ceph-mon[53712]: pgmap v3436: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:41.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:41 vm10.local ceph-mon[53712]: pgmap v3437: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:41.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:41 vm11.local ceph-mon[53973]: pgmap v3437: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:43.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:43 vm10.local ceph-mon[53712]: pgmap v3438: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:43.812 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:43.813 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:43.839 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:43.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:43 vm11.local ceph-mon[53973]: pgmap v3438: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:43.840 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:44.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:44 vm10.local ceph-mon[53712]: pgmap v3439: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:44.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:44 vm11.local ceph-mon[53973]: pgmap v3439: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:46 vm11.local ceph-mon[53973]: pgmap v3440: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:20:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:20:46.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:46 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:20:46.597 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:46 vm10.local ceph-mon[53712]: pgmap v3440: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:46.597 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:20:46.597 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:20:46.597 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:46 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:20:48.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:48 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:20:48.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:48 vm11.local ceph-mon[53973]: pgmap v3441: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:48.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:48 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:20:48.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:48 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:20:48.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:48 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:20:48.591 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:48 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:20:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:48 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:20:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:48 vm10.local ceph-mon[53712]: pgmap v3441: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:48 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:20:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:48 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:20:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:48 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:20:48.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:48 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:20:48.841 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:48.842 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:48.873 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:48.873 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:50.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:50 vm11.local ceph-mon[53973]: pgmap v3442: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:50.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:50 vm10.local ceph-mon[53712]: pgmap v3442: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:53.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:52 vm11.local ceph-mon[53973]: pgmap v3443: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:52 vm10.local ceph-mon[53712]: pgmap v3443: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:53.875 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:53.875 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:53.946 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:53.946 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:54 vm10.local ceph-mon[53712]: pgmap v3444: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:54 vm11.local ceph-mon[53973]: pgmap v3444: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:56 vm10.local ceph-mon[53712]: pgmap v3445: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:56 vm11.local ceph-mon[53973]: pgmap v3445: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:20:58.947 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:20:58.948 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:20:58.974 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:20:58.974 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:20:59.187 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:20:58 vm10.local ceph-mon[53712]: pgmap v3446: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:20:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:20:58 vm11.local ceph-mon[53973]: pgmap v3446: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:00 vm11.local ceph-mon[53973]: pgmap v3447: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:00 vm10.local ceph-mon[53712]: pgmap v3447: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:03 vm11.local ceph-mon[53973]: pgmap v3448: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:03.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:03 vm10.local ceph-mon[53712]: pgmap v3448: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:03.976 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:03.976 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:04.014 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:04.014 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:05 vm11.local ceph-mon[53973]: pgmap v3449: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:05 vm10.local ceph-mon[53712]: pgmap v3449: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:07 vm11.local ceph-mon[53973]: pgmap v3450: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:07 vm10.local ceph-mon[53712]: pgmap v3450: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:09.016 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:09.016 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:09.042 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:09.042 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:09.188 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:09 vm10.local ceph-mon[53712]: pgmap v3451: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:09 vm11.local ceph-mon[53973]: pgmap v3451: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:11 vm11.local ceph-mon[53973]: pgmap v3452: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:11 vm10.local ceph-mon[53712]: pgmap v3452: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:13 vm11.local ceph-mon[53973]: pgmap v3453: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:13 vm10.local ceph-mon[53712]: pgmap v3453: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:14.044 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:14.044 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:14.074 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:14.074 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:15 vm11.local ceph-mon[53973]: pgmap v3454: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:15 vm10.local ceph-mon[53712]: pgmap v3454: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:17 vm11.local ceph-mon[53973]: pgmap v3455: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:17.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:17 vm10.local ceph-mon[53712]: pgmap v3455: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:19.076 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:19.076 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:19.106 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:19.106 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:19.189 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:19 vm10.local ceph-mon[53712]: pgmap v3456: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:19 vm11.local ceph-mon[53973]: pgmap v3456: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:21 vm11.local ceph-mon[53973]: pgmap v3457: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:21 vm10.local ceph-mon[53712]: pgmap v3457: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:23.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:23 vm10.local ceph-mon[53712]: pgmap v3458: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:23 vm11.local ceph-mon[53973]: pgmap v3458: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:24.108 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:24.108 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:24.294 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:24.295 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:25.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:25 vm11.local ceph-mon[53973]: pgmap v3459: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:25.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:25 vm10.local ceph-mon[53712]: pgmap v3459: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:27.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:26 vm11.local ceph-mon[53973]: pgmap v3460: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:27.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:26 vm10.local ceph-mon[53712]: pgmap v3460: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:29.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:28 vm10.local ceph-mon[53712]: pgmap v3461: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:29.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:21:29.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:28 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:21:29.297 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:29.298 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:28 vm11.local ceph-mon[53973]: pgmap v3461: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:29.361 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:21:29.361 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:28 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:21:29.510 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:29.510 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:30 vm11.local ceph-mon[53973]: pgmap v3462: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:30 vm10.local ceph-mon[53712]: pgmap v3462: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:33.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:32 vm11.local ceph-mon[53973]: pgmap v3463: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:33.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:32 vm10.local ceph-mon[53712]: pgmap v3463: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:34.512 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:34.512 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:34.709 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:34.709 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:35.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:35 vm11.local ceph-mon[53973]: pgmap v3464: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:35.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:35 vm10.local ceph-mon[53712]: pgmap v3464: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:37 vm11.local ceph-mon[53973]: pgmap v3465: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:37 vm10.local ceph-mon[53712]: pgmap v3465: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:38.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:38 vm10.local ceph-mon[53712]: pgmap v3466: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:39.090 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:38 vm11.local ceph-mon[53973]: pgmap v3466: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:39.711 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:39.712 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:39.812 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:39.812 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:40 vm11.local ceph-mon[53973]: pgmap v3467: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:40 vm10.local ceph-mon[53712]: pgmap v3467: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:43.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:42 vm10.local ceph-mon[53712]: pgmap v3468: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:42 vm11.local ceph-mon[53973]: pgmap v3468: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:44.814 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:44.814 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:44.844 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:44.844 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:45.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:44 vm10.local ceph-mon[53712]: pgmap v3469: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:44 vm11.local ceph-mon[53973]: pgmap v3469: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:47.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:46 vm10.local ceph-mon[53712]: pgmap v3470: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:46 vm11.local ceph-mon[53973]: pgmap v3470: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:47.934 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:47 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:21:47.935 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:47 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:21:47.935 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:47 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:21:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:47 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:21:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:47 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:21:48.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:47 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:21:49.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:49 vm10.local ceph-mon[53712]: pgmap v3471: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:49.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:49 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:21:49.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:49 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:21:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:49 vm11.local ceph-mon[53973]: pgmap v3471: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:49 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:21:49.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:49 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:21:49.846 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:49.847 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:50.067 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:50.068 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:50.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:50 vm11.local ceph-mon[53973]: pgmap v3472: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:50.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:50 vm10.local ceph-mon[53712]: pgmap v3472: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:53.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:52 vm10.local ceph-mon[53712]: pgmap v3473: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:53.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:52 vm11.local ceph-mon[53973]: pgmap v3473: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:55.069 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:21:55.070 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:21:55.096 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:21:55.097 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:21:55.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:54 vm10.local ceph-mon[53712]: pgmap v3474: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:55.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:54 vm11.local ceph-mon[53973]: pgmap v3474: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:57.184 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:56 vm10.local ceph-mon[53712]: pgmap v3475: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:57.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:56 vm11.local ceph-mon[53973]: pgmap v3475: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:21:59.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:21:58 vm10.local ceph-mon[53712]: pgmap v3476: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:21:59.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:21:58 vm11.local ceph-mon[53973]: pgmap v3476: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:00.098 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:00.098 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:00.126 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:00.127 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:01.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:00 vm11.local ceph-mon[53973]: pgmap v3477: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:01.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:00 vm10.local ceph-mon[53712]: pgmap v3477: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:03.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:03 vm11.local ceph-mon[53973]: pgmap v3478: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:03.436 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:03 vm10.local ceph-mon[53712]: pgmap v3478: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:05.128 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:05.128 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:05.156 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:05.157 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:05.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:05 vm11.local ceph-mon[53973]: pgmap v3479: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:05.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:05 vm10.local ceph-mon[53712]: pgmap v3479: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:07.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:07 vm11.local ceph-mon[53973]: pgmap v3480: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:07.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:07 vm10.local ceph-mon[53712]: pgmap v3480: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:09.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:09 vm10.local ceph-mon[53712]: pgmap v3481: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:09.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:09 vm11.local ceph-mon[53973]: pgmap v3481: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:10.158 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:10.159 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:10.186 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:10.186 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:11.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:11 vm11.local ceph-mon[53973]: pgmap v3482: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:11.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:11 vm10.local ceph-mon[53712]: pgmap v3482: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:13.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:13 vm11.local ceph-mon[53973]: pgmap v3483: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:13.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:13 vm10.local ceph-mon[53712]: pgmap v3483: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:15.188 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:15.188 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:15.214 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:15.215 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:15.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:15 vm11.local ceph-mon[53973]: pgmap v3484: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:15.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:15 vm10.local ceph-mon[53712]: pgmap v3484: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:17.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:17 vm11.local ceph-mon[53973]: pgmap v3485: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:17.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:17 vm10.local ceph-mon[53712]: pgmap v3485: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:19.190 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:19 vm10.local ceph-mon[53712]: pgmap v3486: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:19.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:19 vm11.local ceph-mon[53973]: pgmap v3486: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:20.216 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:20.217 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:20.245 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:20.246 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:21.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:21 vm11.local ceph-mon[53973]: pgmap v3487: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:21.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:21 vm10.local ceph-mon[53712]: pgmap v3487: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:23.435 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:23 vm10.local ceph-mon[53712]: pgmap v3488: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:23.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:23 vm11.local ceph-mon[53973]: pgmap v3488: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:25.247 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:25.248 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:25.274 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:25.274 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:25.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:25 vm11.local ceph-mon[53973]: pgmap v3489: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:25.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:25 vm10.local ceph-mon[53712]: pgmap v3489: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:26.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:26 vm11.local ceph-mon[53973]: pgmap v3490: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:26.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:26 vm10.local ceph-mon[53712]: pgmap v3490: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:29.192 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:29 vm10.local ceph-mon[53712]: pgmap v3491: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:29.193 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:22:29.193 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:29 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:22:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:29 vm11.local ceph-mon[53973]: pgmap v3491: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:22:29.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:29 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm10.qompzp/trash_purge_schedule"}]: dispatch 2026-03-09T19:22:30.276 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:30.276 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:30.480 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:30.480 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:31.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:31 vm11.local ceph-mon[53973]: pgmap v3492: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:31.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:31 vm10.local ceph-mon[53712]: pgmap v3492: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:33.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:33 vm11.local ceph-mon[53973]: pgmap v3493: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:33.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:33 vm10.local ceph-mon[53712]: pgmap v3493: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:35.482 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:35.482 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:35.510 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:35.511 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:35.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:35 vm11.local ceph-mon[53973]: pgmap v3494: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:35.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:35 vm10.local ceph-mon[53712]: pgmap v3494: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:37.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:37 vm11.local ceph-mon[53973]: pgmap v3495: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:37.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:37 vm10.local ceph-mon[53712]: pgmap v3495: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:38.590 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:38 vm11.local ceph-mon[53973]: pgmap v3496: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:38.684 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:38 vm10.local ceph-mon[53712]: pgmap v3496: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:40.512 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:40.513 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:40.540 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:40.540 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:40.840 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:40 vm11.local ceph-mon[53973]: pgmap v3497: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:40.934 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:40 vm10.local ceph-mon[53712]: pgmap v3497: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:43.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:43 vm11.local ceph-mon[53973]: pgmap v3498: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:43.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:43 vm10.local ceph-mon[53712]: pgmap v3498: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:45.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:45 vm11.local ceph-mon[53973]: pgmap v3499: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:45.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:45 vm10.local ceph-mon[53712]: pgmap v3499: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:45.542 INFO:teuthology.orchestra.run.vm10.stderr:++ hostname 2026-03-09T19:22:45.543 INFO:teuthology.orchestra.run.vm10.stderr:+ mount -t nfs vm10.local:/fake /mnt/foo -o sync 2026-03-09T19:22:45.582 INFO:teuthology.orchestra.run.vm10.stderr:mount.nfs: mount system call failed 2026-03-09T19:22:45.582 INFO:teuthology.orchestra.run.vm10.stderr:+ sleep 5 2026-03-09T19:22:47.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:47 vm11.local ceph-mon[53973]: pgmap v3500: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:47.434 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:47 vm10.local ceph-mon[53712]: pgmap v3500: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 170 B/s wr, 0 op/s 2026-03-09T19:22:49.194 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:49 vm10.local ceph-mon[53712]: pgmap v3501: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:49.194 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:49 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:22:49.194 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:49 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:22:49.194 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:49 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:22:49.194 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:49 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:22:49.194 INFO:journalctl@ceph.mon.vm10.vm10.stdout:Mar 09 19:22:49 vm10.local ceph-mon[53712]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:22:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:49 vm11.local ceph-mon[53973]: pgmap v3501: 97 pgs: 97 active+clean; 453 KiB data, 96 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 255 B/s wr, 0 op/s 2026-03-09T19:22:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:49 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:22:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:49 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:22:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:49 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:22:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:49 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:22:49.340 INFO:journalctl@ceph.mon.vm11.vm11.stdout:Mar 09 19:22:49 vm11.local ceph-mon[53973]: from='mgr.14232 192.168.123.110:0/169321951' entity='mgr.vm10.qompzp' 2026-03-09T19:22:50.389 DEBUG:teuthology.exit:Got signal 15; running 1 handler... 2026-03-09T19:22:50.395 DEBUG:teuthology.exit:Finished running handlers